Skip to main content
MedEdPublish logoLink to MedEdPublish
. 2018 Jun 12;7:128. [Version 1] doi: 10.15694/mep.2018.0000128.1

Assessment of Higher Ordered Thinking in Medical Education: Multiple Choice Questions and Modified Essay Questions

Arslaan Javaeed 1,a
PMCID: PMC10699377  PMID: 38074575

Abstract

This article was migrated. The article was marked as recommended.

Background: Multiple choice questions and Modified Essay Questions are two widely used methods of assessment in medical education. There is a lack of substantial evidence whether both forms of questions can assess higher ordered thinking or not.

Objective: The objective of this paper is to assess the ability of a well-constructed Multiple-Choice Question (MCQ) to assess higher ordered thinking skills as compared to a Modified Essay Questions (MEQ) in medical education.

Methods: The medical education literature was searched for articles related to comparison between multiple choice questions and modified essay questions, looking for credible evidence for using multiple choice questions for assessment of higher ordered thinking.

Results and Conclusion: A well-structured MCQ has the capacity to assess higher ordered thinking and because of many other advantages that this format offers. Multiple choice questions should be considered as a preferable choice in undergraduate medical education as literature shows that different levels of Bloom’s taxonomy can be assessed by this assessment format and its use for assessing only lower ordered thinking i.e. recall of knowledge, is not very convincing.

Keywords: multiple choice questions, modified essay questions, assessment higher ordered thinking

Introduction

What is Higher ordered thinking?

Higher ordered thinking is usually defined in reference to the cognitive domain of Bloom’s Taxonomy ( Fig I). First two levels, which are considered as lower ordered thinking, include remembering and understanding whereas rest of the four levels, constituting higher ordered thinking, include application, analysis, evaluation, and creation of knowledge in an ascending order ( Anderson, Lorin, Krathwohll, & Bloom., 2001).

Figure 1. Levels of thinking in revised Bloom's Taxonomy.

Figure 1.

1. Bloom’s Taxonomy to Revised Bloom’s Taxonomy

Bloom’s Taxonomy described and published in 1956 had permeated teaching for almost 45 years before it was modified in 2001. In Table 1, these existing taxonomies of cognition are discussed.

Table 1. Comparison between Bloom's and revised Bloom's Taxonomy.

1.

2. MCQs and MEQ in Medical Education

There has been a considerable revision in undergraduate medical curriculum particularly in the assessment and teaching methodology. Written tests are an essential component of medical education. Objectivity is gradually replacing subjective assessment. Long essay type questions have been substituted by MEQs and MCQs. There is an ongoing debate on which assessment format should be administered to test higher ordered thinking ( Mehta, Bhandari, Sharma, & Kaur, 2016).

Assessment formats are mere tools and their usefulness can be hampered by their poor design, proficiency of its user, deliberate abuse and unintentional misuse (Tom Kubiszyn, 2013). To establish usefulness of a particular assessment format, the following five criteria should be considered: (1) reliability (2) validity (3) influence on future thinking and practice (4) suitability to learners and teachers (5) expenses (to the individual student and institution) ( Vleuten, 1996). Reliability is the degree to which a measurement produces consistent results ( Salkind, 2006) and validity means that how well the test measures which it intends to measure ( American Educational Research Association, American Psychological Association, & National Council on Measurement in Education, 2014).

Discussion

MCQs are extensively used for assessment in medical education owing to their ability to offer a broad range of examination items that incorporate several subject areas. They can be managed in a relatively short period of time. Moreover, they can be marked by a machine which makes the examination standardized ( Epstein, 2007). There is a general perception that MCQs emphasize on knowledge recall i.e. Level I of revised Bloom’s Taxonomy and MEQs are capable of testing higher ordered thinking. The criticism against MCQs is basically due to its poor construction rather than the format itself. A study reveals that in assessing cognitive skills, MCQs significantly correlate with MEQs when their assessment’s content is matched ( Palmer & Devitt, 2007).

The modified essay question is a compromise between an essay and a multiple-choice question. Although it is well documented that a well-constructed MEQ tests higher ordered thinking, it is appropriate to ask if MEQs in undergraduate medical education are well-constructed and test higher ordered thinking. Mostly MEQs only test factual knowledge i.e. lower ordered thinking and at the same time risk significant variation in standards of marking as they are usually hand-marked ( Palmer & Devitt, 2007), rendering it unreasonable as an assessment format for testing a large number of students ( Sam, Hameed, Harris, & Meeran, 2016). Besides, it is a difficult task to construct an MEQ capable of testing higher ordered thinking in students and is more frequently associated with item writing flaws ( Khan & Aljarallah, 2011).

In contrast to MEQ, MCQs are suitable for testing a large number of students as they are machine scored ( Morrison & Walsh Free, 2001). Research shows that multiple choice questions assessing comprehension, application and analysis have been identified. This suggests that the ability of MCQs to assess higher ordered thinking is persistently undervalued and indicates that MCQs have the potential to assess higher ordered thinking ( Scully, 2017). Examples of multiple choice question assessing higher ordered thinking i.e. application ( Table 2), analysis ( Table 3) and evaluation levels ( Table 4, 5 and 6) are as follows:

Table 2. Example of Multiple Choice Question assessing higher ordered thinking i.e. Level III “Application” (Case & Swansin, 2002).

graphic file with name mep-7-19597-g0002.jpg

Table 3. Example of Multiple Choice Question assessing higher ordered thinking i.e. Level IV “Analysis” (Oermann & Gaberson, 2009).

graphic file with name mep-7-19597-g0003.jpg

Table 4. Example of Multiple Choice Question assessing higher ordered thinking i.e. Level V “Evaluation” (Touchie, 2010).

graphic file with name mep-7-19597-g0004.jpg

Table 5. Example of Multiple Choice Question assessing higher ordered thinking i.e. Level V “Evaluation” (Touchie, 2010).

graphic file with name mep-7-19597-g0005.jpg

Table 6. Example of Multiple Choice Question assessing higher ordered thinking i.e. Level IV “Evaluation” (Collins, 2006).

graphic file with name mep-7-19597-g0006.jpg

For a number of purposes, the significance of measuring higher ordered thinking is well renowned in medical education. It has been debated that multiple choice format is useful because it is reliable, objective, unbiased and efficient, cost-effective in nature but incapable of measuring higher ordered thinking. This is not true. A more correct declaration would be that MCQs measuring higher ordered thinking are rarely constructed and MCQs assessing lower ordered thinking are over-presented. One of the reasons of this over presentation is that the most item writers are not formally trained. This emphasizes that format itself is not limited to the assessment of lower ordered thinking. In undergraduate medical education, a well-constructed MCQ can easily assess a student’s ability to apply, evaluate and judge medical education knowledge ( Vanderbilt, Feldman, & Wood, 2013). Nevertheless, writing MCQs capable of assessing higher ordered thinking are challenging ( Bridge, Musial, Frank, Thomas, & Sawilowsky, 2003) but can be developed by following certain guidelines, especially ensuring that item writers are competent in their fields ( Haladyna, & Downing, 2006).

Scully (2017) invalidated the perception that MCQs can only assess lower ordered thinking and Palmer EJ and Devitt (2007) illustrated that the percentage of question testing lower ordered thinking is same in both MCQs and MEQs. It also shows that a well-constructed MCQ is a better tool to assess higher ordered thinking in medical students than an MEQ ( Palmer & Devitt, 2007). There is nothing innate in the MCQ assessment format which prevents testing of higher-ordered thinking ( Norcini, Swanson, Grosso, Shea, & Webster, 1984). Besides, medical schools are training their faculty members to develop multiple-choice questions which ensure assessment of higher ordered thinking of their students. ( Vanderbilt et al., 2013).

Conclusion

The higher ordered thinking in undergraduate medical students can be better assessed through well-constructed multiple-choice questions as compared to modified essay questions. Therefore, well-constructed MCQS should be considered a reasonable substitute for MEQs because of a variety of other advantages it provides over MEQs.

Take Home Messages

A well-constructed MCQS should be considered a reasonable substitute for MEQs because of a variety of other advantages it provides over MEQs.

Notes On Contributors

Dr. Arslaan Javaeed is an assistant professor of Pathology in Poonch Medical College, Rawalakot, Pakistan and is doing his masters in Health Profession Education from Faculty of Education, University of Ottawa, Ottawa, Canada.

Acknowledgments

Thanks are due to Dr. Katherine Moreau for her guidance at every step of this review article.

[version 1; peer review: This article was migrated, the article was marked as recommended]

Declarations

The author has declared that there are no conflicts of interest.

Bibliography/References

  1. American Educational Research Association, American Psychological Association, & National Council on Measurement in Education, & Joint Committee on Standards for Educational and Psychological Testing. (2014). Standards for educational and psychological testing. Washington, DC: AERA; [Google Scholar]
  2. Anderson L. W. & Krathwohl D. R.(2001). A taxonomy for learning, teaching, and assessing: A revision of Bloom’s taxonomy of educational objectives. New York: Longman. [Google Scholar]
  3. Bloom B. S. Engelhart M. D. & Committee of College and University Examiners. (1956). Taxonomy of educational objectives: The classification of educational goals. London: Longmans. [Google Scholar]
  4. Bridge P. D. Musial J. Frank R. Thomas R. & Sawilowsky S.(2003). Measurement practices: Methods for developing content-valid student examinations. Medical Teacher. 10.1080/0142159031000100337 [DOI] [PubMed] [Google Scholar]
  5. Case S. M. & Swanson D. B.(2002). Constructing written test questions for the basic and clinical sciences. Director. 27(21),112. [Google Scholar]
  6. Collins J.(2006). Writing multiple-choice questions for continuing medical education activities and self-assessment modules. RadioGraphics. 26(2),543–551. 10.1148/rg.262055145 [DOI] [PubMed] [Google Scholar]
  7. Epstein R. M.(2007). Medical education - Assessment in medical education. New England Journal of Medicine. 356(4),387–396. 10.1056/NEJMra054784 [DOI] [PubMed] [Google Scholar]
  8. Khan Z. & Aljarallah B. M.(2011). Evaluation of modified essay questions and multiple choice questions as a tool for assessing the cognitive skills of undergraduate medical students. International Journal of Health Sciences. 5(1),39–43. [PMC free article] [PubMed] [Google Scholar]
  9. Mehta B. Bhandari B. Sharma P. & Kaur R.(2016). Short answer open-ended versus multiple-choice questions : A comparison of objectivity. 52(3),173–182. [Google Scholar]
  10. Morrison S. & Walsh Free K. W.(2001). Writing multiple-choice test items that promote and measure critical thinking. The Journal of Nursing Education. 40(1),17–24. [DOI] [PubMed] [Google Scholar]
  11. Norcini J. J. Swanson D. B. Grosso L. J. Shea J. a & Webster G. D.(1984). A comparison of knowledge, synthesis, and clinical judgment. Multiple-choice questions in the assessment of physician competence. Evaluation & the Health Professions. 7,485–499. 10.1177/016327878400700409 [DOI] [PubMed] [Google Scholar]
  12. Oermann M. H. & Gaberson K. B.(2009). Evaluation and testing in nursing education. New York, Springer. [Google Scholar]
  13. Palmer E. J. & Devitt P. G.(2007). Assessment of higher order cognitive skills in undergraduate education: modified essay or multiple choice questions? Research paper. BMC Medical Education. 7(1),49. 10.1186/1472-6920-7-49 [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Čapek Radan Peter McLeod C. C. & B. J. R.(2005). Do accompanying clinical vignettes Improve student scores on multiple choice questions testing factual knowledge? Medical Science Educator. 20(2),110–119. [Google Scholar]
  15. Salkind N. J.(2006). In Tests & measurement for people who (think they) hate tests & measurement. Thousand Oaks, Calif: SAGE Publications. [Google Scholar]
  16. Sam A. H. Hameed S. Harris J. & Meeran K.(2016). Validity of very short answer versus single best answer questions for undergraduate assessment. BMC Medical Education. 16(1),266. 10.1186/s12909-016-0793-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Scully D.(2017). Constructing multiple-choice items to measure higher-order thinking. Practical Assessment, Research & Evaluation. 22(4),1–13. [Google Scholar]
  18. Haladyna T. M. & Downing S. M.(2006). Handbook of test development. Mahwah, N.J: L. Erlbaum. [Google Scholar]
  19. Tiemeier A. M. Stacy Z. A. & Burke J. M.(2011). Innovations in pharmacy using multiple choice questions written at various Bloom’s Taxonomy levels to evaluate student performance across a therapeutics sequence. INNOVATIONS. 2(2). [Google Scholar]
  20. Kubiszyn T. & Borich G. D.(2013). Educational testing and measurement: Classroom application and practice. New York, NY: Wiley. [Google Scholar]
  21. Touchie C.(2010). Medical Council of Canada Guidelines for the Development of Multiple-Choice Questions. [Google Scholar]
  22. Vanderbilt A. A. Feldman M. & Wood I. K.(2013). Assessment in undergraduate medical education: a review of course exams. Medical Education Online. 18(1),20438. 10.3402/meo.v18i0.20438 [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Vleuten C. Van Der.(1996). The assessment of professional competence: developments, research and practical implications. Advances in Health Sciences Education. 1,41–67. 10.1007/BF00596229 [DOI] [PubMed] [Google Scholar]
MedEdPublish (2016). 2019 Mar 13. doi: 10.21956/mep.19597.r29102

Reviewer response for version 1

sathyanarayanan varadarajan 1

This review has been migrated. The reviewer awarded 4 stars out of 5 This interesting article is about assessing the ability of a well-constructed Multiple-Choice Question (MCQ) to assess higher order thinking skills as compared to a Modified Essay Questions (MEQ) in medical education. This has been done by the author by searching and analyzing the medical education literature available. The author has compared MEQ and MCQ with suitable Examples of multiple choice questions assessing higher order thinking i.e. application, analysis and evaluation levels. The author concludes that well-constructed MCQs should be considered as a reasonable substitute for MEQs because of a variety of advantages it provides over MEQs. But the real challenge lies in constructing MCQs appropriately that can assess the higher order thinking skills truly and effectively. This paper will be useful for all medical educators, who are involved in Evaluation and Assessment.

Reviewer Expertise:

NA

No decision status is available

MedEdPublish (2016). 2018 Jun 24. doi: 10.21956/mep.19597.r29103

Reviewer response for version 1

Sateesh Babu Arja 1

This review has been migrated. The reviewer awarded 4 stars out of 5 I enjoyed reading this paper. It is an interesting paper which is in favor of multiple choice questions over modified essay questions. I agree with authors it is required to see how well constructed is the assessment rather than the format of the assessment. As long as the assessments fulfill the important traits like Validity, Reliability, feasibility, educational effect, and catalytic effect, they are good practices. Validity can be determined by the Miller's pyramid, the content of the exam, and mapping with learning objectives. It is known that MCQs are clearly helpful in assessing application and analysis levels. But it is interesting to see that MCQs can be used to assess evaluation level too. This paper is good for faculty members, academic administrators, and assessment committees.

Reviewer Expertise:

NA

No decision status is available

MedEdPublish (2016). 2018 Jun 17. doi: 10.21956/mep.19597.r29100

Reviewer response for version 1

P Ravi Shankar 1

This review has been migrated. The reviewer awarded 4 stars out of 5 The author discusses an important point in this article. He describes how a well-constructed multiple choice question (MCQ) can test higher-order skills. The author has provided some examples of MCQs testing higher order skills and has described both the original Bloom’s taxonomy and its recent revision. The article could be further strengthened by the author providing tips to construct MCQs testing higher order skills. In offshore Caribbean medical schools MCQs are widely used for assessment and are modeled after the United States Medical Licensing exam (USMLE) pattern. A problem is MCQs are the predominant method of assessment and other methods are not widely used. Many of these schools concentrate on preparing students for success in the USMLE which predominantly uses single response MCQs. MCQs are objective, unbiased and can be evaluated by machine and are widely used in various licensing and other high stakes exams. All medical educators will find this article to be of interest.

Reviewer Expertise:

NA

No decision status is available

MedEdPublish (2016). 2018 Jun 17. doi: 10.21956/mep.19597.r29101

Reviewer response for version 1

Tayyaba Khan 1

This review has been migrated. The reviewer awarded 5 stars out of 5 I thoroughly enjoyed reading this research article. Mostly multiple choice questions are used to assess factual knowledge only and this article indicates need of wellconstructed mcqs for their proper usage.

Reviewer Expertise:

NA

No decision status is available


Articles from MedEdPublish are provided here courtesy of Association for Medical Education in Europe

RESOURCES