Skip to main content
MedEdPublish logoLink to MedEdPublish
. 2021 Feb 5;10:37. [Version 1] doi: 10.15694/mep.2021.000037.1

Questioning the questions: Methods used by medical schools to review internal assessment items

Bindu Menon 1, Jolene Miller 1, Lori M DeShetler 1,a
PMCID: PMC10939609  PMID: 38486513

Abstract

This article was migrated. The article was marked as recommended.

Objective: Review of assessment questions to ensure quality is critical to properly assess student performance. The purpose of this study was to identify processes used by medical schools to review questions used in internal assessments.

Methods: The authors recruited professionals involved with the writing and/or review of questions for their medical school’s internal assessments to participate in this study. The survey was administered electronically via an anonymous link, and participation was solicited through the DR-ED listserv, an electronic discussion group for medical educators. Responses were collected over a two-week period, and one reminder was sent to increase the response rate. The instrument was comprised of one demographic question, two closed-ended questions, and two open-ended questions.

Results: Thirty-nine respondents completed the survey in which 22 provided the name of their institution/medical school. Of those who self-identified, no two respondents appeared to be from the same institution, and participants represented institutions from across the United States with two from other countries. The majority (n=32, 82%) of respondents indicated they had a process to review student assessment questions. Most participants reported that faculty and course/block directors had responsibility for review of assessment questions, while some indicated they had a committee or group of faculty who was responsible for review. Most focused equally on content/accuracy, formatting, and grammar as reported. Over 81% (n=22) of respondents indicated they used NBME resources to guide review, and less than 19% (n=5) utilized internally developed writing guides.

Conclusions: Results of this study identified that medical schools are using a wide range of item review strategies and use a variety of tools to guide their review. These results will give insight to other medical schools who do not have processes in place to review assessment questions or who are looking to expand upon current procedures.

Keywords: Medical education, Assessment, Test item review

Introduction

It is widely acknowledged that well-designed assessments positively impact student learning and drive the robust growth of a curriculum by identifying curricular strengths and weaknesses ( Norcini et al., 2011). Medical schools have long recognized and emphasized the importance of internal examinations in ensuring that the graduating students are equipped with the knowledge and skills required to be competent and safe medical practitioners ( Miller, 1990). Well-written tests benefit both students and faculty. They motivate student learning and provide students with accurate performance feedback. These tests benefit faculty by providing feedback on teaching effectiveness. Conversely, the detrimental effects of poor item quality have also been well recognized by Downing (2005) and Tarrant and Ware (2008). Past research ( Downing, 2005; Jozefowicz et al., 2002; Rodriguez-Diez et al., 2016) has shown that multiple-choice questions often contain flaws that contribute to measurement error. Item-writing flaws have been shown to lead to construct-irrelevant variance thereby affecting the pass-fail outcomes for students in previous studies Downing, 2005; Downing, 2002). Generating quality assessments with well-written items on a regular basis has been reported a challenge by several medical schools according to Case, Holtzman and Ripkey (2001) and Pinjani, Umer and Sadaf (2015).

During the 2018-19 academic year, medical student feedback from course evaluations at our institution consistently identified issues with internal assessments. The identified problems included typographical, grammatical, and formatting errors as well as unclear question stems. Faculty were responsible for writing assessment questions, and course directors were charged with developing the assessments, but our medical school did not have a systematic process in place to review each assessment question prior to use in internal examinations. In the fall of 2019, college leadership established an item review committee to address student concerns by establishing a process for peer and editorial review of assessment items. Membership on this committee included faculty representing different areas of expertise: item writing, assessment, content, and editing. Soon after convening, the committee recognized the need for not only review of each assessment question, but also the need for a guide to aid faculty and directors in writing quality assessments. During committee review, members check each question’s formatting, grammar, and structure. If issues about the content of the item, such as questionable accuracy or confusing presentation, are identified, the course director is notified.

Peer-review of assessment questions for writing flaws is an effective way to improve question quality and performance ( Abozaid, Park and Tekian, 2017; Malua-Aduli and Zimitat, 2012; Wallach et al., 2006). To assist the committee in its work, we were interested in how other medical schools reviewed assessment questions but were unable to find any research regarding the issue. The purpose of this study was to determine what processes, if any, medical schools use to review test items before the items are used on student assessments. We specifically sought to understand which individuals and groups were involved in review processes and what they included in their review. This purpose was achieved by answering the following research question: What methods do medical schools use to review questions that will be used to assess students’ knowledge and competence in internal examinations?

Methods

Design

We used a descriptive study with an online questionnaire to identify if medical schools have processes to review assessment items and to determine what methods they use in the review of questions. The Assessment Item Review survey (Supplementary File 1) consisted of one demographic question, two closed-ended questions, and two open-ended questions. The research was reviewed by The University of Toledo Social, Behavioral, and Educational Institutional Review Board and was found that the study did not meet the definition of human subjects’ research as outlined in 45 CFR 46.102(e)(1), and therefore did not require Institutional Review Board oversight or approval. We recruited medical school professionals into the study by email during spring 2020. The purpose of this study was explained, and participants were provided with an anonymous link to take the survey. Completion of the survey constituted informed consent.

Sample

The sample was solicited from professionals subscribed to the DR-ED listserv, an electronic discussion group for medical educators. This email distribution list was selected because the membership includes medical school professionals who are involved with student assessment.

Outcome measures

The survey contained an optional demographic question in which participants were asked to provide their institution/name of medical school. Two closed-ended questions followed. Participants were asked to indicate whether they had a process to review student assessment questions before they are used. If “No” were selected, the respondent was taken to the last question in the survey. The second closed-ended item asked participants to select which people or groups review student assessment questions before they are used, and what aspect(s) of questions they review. Respondents could select all that apply. Options for individuals and groups included Faculty member writing the question, Group of faculty members teaching related topics, Unit (course/block) director(s), Non-faculty academic staff/coordinator(s), Assessment question review committee, Curriculum committee, and Other. For the aspects of questions each individual/group reviews, response options were Content/accuracy, Item formatting, Grammar/spelling, and Other.

Two open-ended questions followed. Respondents were asked to list all sources and documents their medical school uses to guide student assessment question review (e.g., National Board of Medical Examiners [NBME] item writing manual, internally developed writing guide, NBME laboratory values). The last question of the survey prompted participants to share any other useful information regarding their medical school’s assessment question review process.

Analysis

The analysis involved comparing the self-identified respondents’ institution to determine the possibility of duplication of responses from the same medical school. Next, tallies were run for the first closed-ended question to calculate the percentage of respondents who had a process for reviewing assessment items. In the second closed-ended question, we analyzed the frequency of people and groups that were selected for reviewing assessment questions, and the frequency for the type of review was analyzed to understand the roles of the people and groups tasked with reviewing assessment questions.

The second part of the analysis included coding of the qualitative responses. From the first open-ended question pertaining to sources and documents that the participants’ medical school uses to guide assessment question review, we grouped common terms and ranked sources from most to least cited. A qualitative analysis was also conducted on the last question regarding other useful information that participants chose to share, and themes were created based on their responses. A frequency threshold of 15% was utilized for identifying themes in the open-ended responses.

Results/Analysis

A total of 39 participants completed the survey. Of this total, 22 provided the name of their institution/medical school. For those who self-identified, no two respondents appeared to be from the same institution, and participants represented schools from across the United States with two from other countries. All 39 participants answered the question about whether their school had a process to review student assessment questions. Just over 82% (n=32) reported that their medical school did have a process.

Table 1 shows the frequency of which person or group reviews assessment questions at the participants’ medical school, and of which aspect(s) the review consists (e.g., grammar/spelling).

Table 1. Person or group who reviews student assessment questions and aspects of the review.

Person or Group Aspect of Review
Content/Accuracy(number of respondents) Item Formatting(number of respondents) Grammar/Spelling(number of respondents) Other(number of respondents)
Faculty member writing the question 26 18 20 4
Group of faculty members teaching related topics 11 10 9 1
Unit (course/block) director(s) 21 21 20 6
Non-faculty academic staff/coordinator(s) 0 9 10 2
Assessment question review committee 12 13 13 5
Curriculum committee 2 0 0 0
Other (list) a 2 3 3 1
a

Medical Education Center, Academic Deans, Director of Assessment/Assistant Dean of assessment

The most common response (n=26) was that the faculty member writing the question holds responsibility for the content/accuracy of the assessment question. The next highest frequency (n=21) was unit (course/block) directors for the review of both content/accuracy and item formatting. Close behind, 20 participants indicated that the faculty member writing the question reviews for grammar/spelling, and 20 also reported that the unit directors review grammar/spelling. Less than half (n=18, 46%) of the respondents indicated that the faculty member writing the question at their medical school was responsible for item formatting. All other frequencies for the remaining choices of people and groups by review task were one-third or less. From these results, faculty and unit directors shared the highest frequency for review of assessment questions followed by assessment question review committees.

The type of question review was evenly dispersed among content/accuracy, item formatting, and grammar/spelling. The “Other” category was rarely chosen. The task of reviewing for content/accuracy was reported most (n=26) for the faculty member writing the question. Respondents indicated that item formatting was most carried out by unit directors (n=21). Meanwhile, grammar/spelling was selected as the responsibility of both faculty members writing the question and unit directors by 20 participants. Nine respondents indicated that non-faculty academic staff/coordinators reviewed item formatting, and 10 reported that they reviewed grammar/spelling; however, none of the participants selected content/accuracy for non-faculty academic staff/coordinators. By question review task, the faculty member writing the question and unit directors were the highest frequency. Only two medical schools indicated that their curriculum committee was involved with the question review process.

Twenty-seven participants provided sources and documents that their medical school uses to guide student assessment question review. The majority (n=21, 78%) of respondents listed the NBME item writing guide as a source they use to guide question review. Almost a third (n=8, 30%) of participants included NBME laboratory values as a document they utilize in the review process. Internally developed writing guides and item writing courses/workshops were each listed by five respondents.

Additional comments were provided by 20 participants. Three responses centered on the implication for faculty training to facilitate item review. For example, one respondent stated, “It is very important that teachers take a training course in learning assessment.” Another indicated that item review is best handled by course faculty, but individual faculty may view the process as a “waste of time.”

There were six comments regarding the quality of test questions. One participant explained, “Having a quality item bank software and good quality items that were peer reviewed before they were permitted to be used...were really important.” Another respondent described his/her review process in which item quality is reviewed and verified to confirm the quality of questions. Some who discussed the quality also included terms for validating their questions.

A third theme that emerged from 30% (n=6) of the comments was related to the roles of block/course directors in test item review. One participant stated that they have three levels of review, one of which includes the course director. Similarly, another participant said, “We have assessment vetting sessions by block directors.” Another indicated that following exam item review, suggestions are provided to course directors who then share feedback with the faculty.

Over one-third (n=7) of the comments focused on test item performance. Various respondents provided information regarding how their medical school tracks and uses item performance. For example, one respondent stated, “The performance statistics are used to update/improve question stems and answer choices.” Likewise, another said, “We track item performance before/after committee review.” Others noted tracking item performance over time or using statistical analytics for quality improvement. It should be noted that of the seven who indicated that they did not have a review process prior to items being used on an assessment, two shared that they analyzed item performance statistics after items are used.

Lastly, 75% (n=15) of respondents provided comments on the responsibility for test question review. Two respondents discussed a team approach, while another indicated that his/her medical school utilizes a peer review process. One participant said, “questions are viewed by at least two other faculty.” As mentioned previously, several made references to block directors, who held responsibility for item review at their schools. It appeared that some institutions split the responsibility of test question review among multiple groups (e.g., Assessment Office, Item Review Committee, and Course Director), and one had different processes depending on the medical student year (MD1 versus MD2).

Discussion

Most of the participating medical schools had a process to review assessment questions before they are used on examinations. The responsibility for and the focus of the review differed by institution. We found that faculty and directors were most often responsible for the review of assessment questions based on these data. Assessment question review committees, while established at some of the respondents’ medical schools, were not as commonly reported as oversight for the review process as these individuals. In fact, it can be inferred from the data that only one-third of respondents had an assessment question review committee.

Because of the importance of internal examinations to assess student knowledge and competence, the greatest concern with poorly written items is construct-irrelevant variance. This is variance in examination scores that has nothing to do with student knowledge and competence. While there are a number of factors that contribute to this variance ( Downing, 2002), technical flaws in items contribute to irrelevant difficulty and testwiseness ( Paniagua and Swygert, 2016). Although examination questions are expected to vary in difficulty, that difficulty should be based on the content being assessed, not the structure of the question. The NBME has highlighted issues that contribute to irrelevant difficulty such as numerical responses presented in an illogical order, and the response option “None of the above.” Irrelevant difficulty introduces measurement error that decreases student scores, while testwiseness increases the scores for students who know how to take tests. These sorts of flaws include grammatical or logical cues (allowing the testwise student to rule out one or more options) and correct responses that are different in terms of length and detail ( Paniagua and Swygert, 2016).

The item review committee members in our medical school soon became cognizant of the fact that in ensuring test quality, the ultimate onus is on the faculty who are also the content experts, with the committee providing a more editorial review. Developing valid and reliable test items without construct-irrelevant variance is a critical skill for the faculty to hone. It seems institutions are giving more attention to faculty development to improve the quality of their exams ( Jozefowicz et al., 2002; Abdulghani et al., 2015; AlFaris et al., 2015; Iramaneerat, 2012; Naeem, van der Vleuten and AlFaris, 2012), as studies have shown that faculty development and providing training in exam item writing improves the process of item writing and quality of exams ( Naeem, van der Vleuten and AlFaris, 2012; Tunk, 2001; Kim et al., 2010). The importance of faculty development was reflected in respondents’ comments.

In institutions where individual faculty are solely responsible for the quality of assessment items, the use of performance analytics could be one way of tracking student progress and reviewing item performance. Yet, it is ideal to assign oversight of the items to a committee or director to ensure the overall quality of the exam, particularly in areas such as grammar and formatting. The establishment of an item review committee in our medical school that oversees all the test items to ensure uniformity and flow of reading has reduced the stress typically caused by these types of flaws as evident in medical student feedback.

The majority of the respondents (n=21) reported use of the NBME item writing guide to facilitate review of their assessments. While the NBME guide is a comprehensive document that details several methods to avoid issues such as construct-irrelevant variance, there are other issues that may appear in exams that create unnecessary stress to the exam takers. To address these problems, the item review committee from our medical school developed an internal style guide (Supplementary File 2) to direct the faculty writing questions and to guide the committee’s review. The style guide, while maintaining the major directives in the NBME guide, includes pointers for writers to ensure ease of reading and uniformity of the questions. The style guide includes recommendations for uniformity of units, drug names, etc., and emphasizes proper placement and style of tables and figures in the question stem. The brevity of our style guide (11 pages compared to the NBME guide’s 84 pages) allows it to serve as a quick reference. This internal guide was endorsed by the curriculum committee and disseminated to faculty to encourage use and improve test-item quality. The student feedback on assessments suggests a positive response thus far, and our item review committee plans to analyze these data after one full cycle.

Limitations

One limitation to this study was the number of responses. We anticipated a higher response rate because the listserv used for solicitation is widely used by professionals in medical education worldwide. A reason for low participation could be due to individuals choosing not to participate if their institution did not have a formal item review process in place. Related to this may have been a misunderstanding of the phrase “process to review student assessment questions before they are used.” For example, if the faculty member writing the item is responsible for review, would a potential respondent consider that to be something other than a review process and decline to participate in the research? In addition, the administration of the survey coincided with the early stages of the COVID-19 pandemic during which faculty and administrators were occupied with higher priorities.

Future Implications

The current study shows that one third of survey participants reported the existence of a similar committee to ensure exam quality in their institutions ( Table 1). This practice may be in development at other medical schools, and hence we feel it is worthwhile to conduct another study to investigate the function and effectiveness of item review committees. What are best practices for use of such a committee with respect to item writers’ and course directors’ review? What is the appropriate combination of skills needed by members of the committee? It would also be of interest to compare our medical school style guide with the internally developed guides from other institutions in order to identify key components of these documents.

Conclusion

This study provides valuable information about the practices employed by various medical schools in ensuring assessment quality. The results identified that medical schools are using a wide range of practices to ensure assessment quality. The diversity of item review strategies, from no formal review process to multi-step processes, in combination with a variety of tools to guide their review, highlights the need for medical schools to develop item review processes that reflect their resources, needs, and culture. The survey results will be helpful for institutional authorities planning to adopt new processes to review assessment questions or looking to expand upon current procedures.

Take Home Messages

  • Most participating institutions had a process to review assessment questions before use, which suggests that assessment item review is considered best practice.

  • Faculty development on exam item writing improves the process of question creation and exam quality.

  • Assign oversight of assessment items to a committee or director to ensure the overall exam quality, particularly in areas such as grammar and formatting.

  • Membership on an item review committee should include one or more non-medical educators with grammar and editing skills.

  • Use of the NBME item writing guide or an internally developed writing guide is helpful in facilitating review of assessment items.

Notes On Contributors

Lori M. DeShetler, PhD, is the Assistant Dean for Assessment and Accreditation in the Department of Medical Education at The University of Toledo. Dr. DeShetler’s current research interests include assessment, COVID-19 impact on medical students, curriculum mapping, and implications of the opioid crisis on medical education. ORCID ID: https://orcid.org/0000-0002-4566-7111

Bindu Menon, PhD, is Assistant Professor in the Department of Medical Education at The University of Toledo. Dr. Menon’s research interests include vertical integration of foundational science elements into clinical years, COVID-19 impact on medical students, cognitive assessment of student mastery in assessments and analysis of NBME subject examination results. ORCID ID: https://orcid.org/0000-0002-4436-8208

Jolene M. Miller, MLS, is the Director of the Mulford Health Science Library at The University of Toledo. Ms. Miller serves on the university’s MD Program Item Review Committee. Her research interests are the use of reflective practice by health science librarians and role of emotion regulation in library administrators. ORCID ID: https://orcid.org/0000-0003-4422-2708

Declarations

The author has declared that there are no conflicts of interest.

Ethics Statement

This research was reviewed by The University of Toledo Social, Behavioral, and Educational Institutional Review Board and was found that the study did not meet the definition of human subjects’ research as outlined in 45 CFR 46.102(e)(1), and therefore did not require Institutional Review Board oversight or approval.

External Funding

This article has not had any External Funding

Acknowledgments

We thank The University of Toledo medical school’s Item Review Committee for their support in this study and for permitting the publication of all information and documentation published with this article.

Previous Presentations: None.

[version 1; peer review: This article was migrated, the article was marked as recommended]

Bibliography/References

  1. Abdulghani H. M., Ahmad F., Irshad M., Khalil M. S., et al. (2015) Faculty development programs improve the quality of multiple choice questions items’ writing. Scientific Reports. 5, p.9556. 10.1038/srep09556 [DOI] [PMC free article] [PubMed] [Google Scholar]
  2. Abozaid H. Park Y. S. and Tekian A.(2017) Peer review improves psychometric characteristics of multiple choice questions. Medical Teacher. 39(1), pp.S50–S54. 10.1080/0142159X.2016.1254743 [DOI] [PubMed] [Google Scholar]
  3. AlFaris E., Naeem N., Irfan F., Qureshi R., et al. (2015) A one-day dental faculty workshop in writing multiple choice questions: An impact evaluation. Journal of Dental Education. 79(11), pp.1305–1313. [PubMed] [Google Scholar]
  4. Case S. M. Holtzman K. and Ripkey D. R.(2001) Developing an item pool for CBT: a practical comparison of three models of item writing. Academic Medicine. 76(10), pp.S111–S113. 10.1097/00001888-200110001-00037 [DOI] [PubMed] [Google Scholar]
  5. Downing S. M.(2002) Construct-irrelevant variance and flawed test questions: Do multiple-choice item-writing principles make any difference? Academic Medicine. 77(10), pp.S103–S104. 10.1097/00001888-200210001-00032 [DOI] [PubMed] [Google Scholar]
  6. Downing S. M.(2005) The effects of violating standard item writing principles on tests and students: The consequences of using flawed test items on achievement examinations in medical education. Advances in Health Sciences Education. 10(2), pp.133–143. 10.1007/s10459-004-4019-5 [DOI] [PubMed] [Google Scholar]
  7. Iramaneerat C.(2012) The impact of item writer training on item statistics of multiple-choice items for medical student examination. Siriraj Medical Journal. 64, pp.178–182. [Google Scholar]
  8. Jozefowicz R. F., Koeppen B. M., Case S., Galbraith R., et al. (2002) The quality of in-house medical school examinations. Academic Medicine. 77(2), pp.156–161. 10.1097/00001888-200202000-00016 [DOI] [PubMed] [Google Scholar]
  9. Kim J., Chi Y., Huensch A., Jun H., et al. (2010) A case study on an item writing process: Use of test specifications, nature of group dynamics, and individual item writers’ characteristics. Language Assessment Quarterly. 7(2), pp.160–174. 10.1080/15434300903473989 [DOI] [Google Scholar]
  10. Malua-Aduli B. S. and Zimitat C.(2012) Peer review improves the quality of MCQ examinations. Assessment & Evaluation in Higher Education. 37(8), pp.919–931. 10.1080/02602938.2011.586991 [DOI] [Google Scholar]
  11. Miller G. E.(1990) The assessment of clinical skills/competence/performance. Academic Medicine. 65(9), pp.S63–S67. 10.1097/00001888-199009000-00045 [DOI] [PubMed] [Google Scholar]
  12. Naeem N. van der Vleuten C. and AlFaris E. A.(2012) Faculty development on item writing substantially improves item quality. Advances in Health Sciences Education. 17, pp.369–376. 10.1007/s10459-011-9315-2 [DOI] [PubMed] [Google Scholar]
  13. Norcini J., Anderson B., Bollela V., Burch V., et al. (2011) Criteria for good assessment: Consensus statement and recommendations from the Ottawa 2010 conference. Medical Teacher. 33(3), pp.206–214. 10.3109/0142159X.2011.551559 [DOI] [PubMed] [Google Scholar]
  14. Paniagua M. and Swygert K.(2016) Constructing Written Test Questions for the Basic and Clinical Sciences. 4th ed. Philadelphia, PA: National Board of Medical Examiners. [Google Scholar]
  15. Pinjani S. Umer M. and Sadaf S.(2015) Faculty engagement in developing an internship entry test. Medical Education. 49(5), pp.540–541. 10.1111/medu.12721 [DOI] [PubMed] [Google Scholar]
  16. Rodriguez-Diez M. C., Alegre M., Diez N., Arbea L., et al. (2016) Technical flaws in multiple-choice questions in the access exam to medical specialties ("examen MIR") in Spain (2009-2013). BMC Medical Education. 16(47). 10.1186/s12909-016-0559-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Tarrant M. and Ware J.(2008) Impact of item-writing flaws in multiple-choice questions on student achievement in high-stakes nursing assessments. Medical Education. 42(2), pp.198–206. 10.1111/j.1365-2923.2007.02957.x [DOI] [PubMed] [Google Scholar]
  18. Tunk J.(2001) The effect of training on test item writing on test performance of junior high students. Educational Studies. 27(2), pp.129–142. 10.1080/03055690120050374 [DOI] [Google Scholar]
  19. Wallach P. M., Crespo L. M., Holtzman K. Z, Galbraith R. M., et al. (2006). Use of a committee review process to improve the quality of course examinations. Advances in Health Sciences Education. 11(1), pp.61–68. 10.1007/s10459-004-7515-8 [DOI] [PubMed] [Google Scholar]
MedEdPublish (2016). 2021 May 30. doi: 10.21956/mep.18904.r27012

Reviewer response for version 1

Ken Masters 1

This review has been migrated. The reviewer awarded 3 stars out of 5 An interesting study on the methods used by medical schools to review assessment questions to ensure quality.I am pleased that the authors have supplied a copy of their questionnaire and documentation for schools intending to introduce some form of question quality review.For me, perhaps the most eye-opening part of this paper is that there are medical schools that do not have a system to review exam questions. It is pleasing to see that this particular school recognised the need to implement such a process. As the authors note, a weakness of the paper is that many possible participants might not have participated because their institution does not have such a process. So, while the paper does give some insight into those processes followed, it does not give an indication of how widespread they are. It would be useful to have follow-up research that addresses this question.Some other issues:• “A frequency threshold of 15% was utilized for identifying themes in the open-ended responses.” This does appear to be rather arbitrary, and I’m not entirely sure that setting such a threshold has support from the literature.• “from across the United States with two from other countries”. I think it would be a good idea to name those other countries, rather than have them classified as “other”.• It is not clear if the survey form was mailed to the participants (and then returned), or if the participants completed the form online (e.g. through Google forms or similar).I look forward to Version 2 of the paper in which these issues and those raised by the first reviewer are addressed.

Reviewer Expertise:

NA

No decision status is available

MedEdPublish (2016). 2021 May 19. doi: 10.21956/mep.18904.r27011

Reviewer response for version 1

Keith Wilson 1

This review has been migrated. The reviewer awarded 4 stars out of 5 The present study aimed to ascertain methods in use by medical schools in their vetting of questions used on internal examinations. The authors were exploring options for improving question quality in response to assessment feedback from students. They emphasize the need to develop assessments that assess content rather than influences of construct-irrelevant variance. The study was a survey of medical educators that included those that were involved in assessment. They chose a convenience sample derived from members of the DR-ED listserv. Unfortunately, their response rate was lower than expected and they discussed possible reasons for this. Additionally, it is unclear who were the respondents – e.g. general faculty, committee chairs.The questionnaire was brief and explored whether institutions had a process to review questions before they appeared on an examination. It would have been helpful to know whether these same institutions had a post-exam review and how they incorporated a quality improvement cycle. The authors performed a qualitative analysis of the open-ended responses and coded these although it is unclear from the text what method of qualitative analysis was used.Respondents highlighted that much of the onus for content/accuracy fell to the question writers themselves and/or course/unit directors. The authors noted that faculty development would be key to ensuring consistency between writers. Personally, I think a more systematic approach is warranted and indeed some of the respondents had processes in place that involved centralized oversight/quality measures.Despite the lower-than-expected response rate, this is a good article for those that are embarking on improving their test banks. The authors make the case for a more systematic approach in reviewing test items. Although the respondents were mostly from US schools, the concepts translate well to medical schools around the world. Additionally, they included in a second supplementary file their summarized recommendations for question writing: although there are specifics to their school, this resource could be adapted to suit other institutions as it contains many helpful tips.

Reviewer Expertise:

NA

No decision status is available


Articles from MedEdPublish are provided here courtesy of Association for Medical Education in Europe

RESOURCES