In a previous issue of Canadian Family Physician, Desrosiers mentioned that “in the process of becoming physicians, we also become teachers.”1 One might add that in becoming teachers, we also become evaluators! In fact, learning assessment is at the core of the learning process. There are few pedagogic research results that garner as much consensus as the importance of assessment in the learning context.2-4 In this article, I present 5 learning assessment strategies that physician teachers and evaluators could leverage in their teaching, be it in a preclerkship, clerkship, or residency setting.
How to render the invisible visible?
The concern with the assessment process is that learning is, alas, invisible. We could never accurately visualize student and resident learning. Therefore, the challenge is to implement a system that renders learning visible, allowing us to measure or observe it to ensure the validity and fidelity of the inferences. Laurier et al5 place assessment within a dynamic of change; learning represents the result located between a final situation and an initial situation. Within a competency-based framework, it is possible to implement a number of strategies that allow us to infer the competency being assessed based on a set of performances. The key is to strategically choose performances, such as answering questions on an examination, discussing case studies, or conducting physical examinations, that would allow evaluators to make informed decisions on the quality of the progression of learning.
Keep the assessment contract in mind
The concept of the assessment contract is borrowed from the didactic contract developed by Brousseau.6 Generally speaking, learners and teachers negotiate, more or less explicitly, the elements of the assessment. This can be illustrated by the propensity of some learners, knowing they are undergoing a summative assessment, to adopt behaviour that emphasizes the best aspect of their competencies while masking the less favourable aspects. Another classic example is the fact that learners will study and primarily focus on learning that is assessed, to the detriment of learning that is not. All learning is important, but that which is assessed and could impact their progression is considered more important!
Assessing means respecting a rigorous process
For many people, assessments are reduced to a mark (eg, 90%) or a grade (eg, pass). These aspects are important because they are effectively a part of the assessment approach and represent the most visible component of it. Yet, to arrive at a decision, assessing physicians must implement an approach that is formal or informal, instrumented or not, and planned or not. In all cases, the assessment will consist of the following dimensions: planning, information gathering, the judgment and the decision, and finally, communication of the result. If the approach is formal, instrumented, and planned, as is often the case in preclerkships, the dimensions will be easily observable. Learners will be informed of the examination sessions (planning). Following the administration of the examination, teachers will grade the examination, or it will be graded automatically (information gathering). Once information is revealed through the grading process, a judgment will be made followed by a decision (eg, learners who have not passed the threshold score must retake the examination). Finally, there is the communication of the decision (eg, communicating the grade to the learner or informing them of the need to retake the examination).
The situation I have just described is characteristic of a summative approach. Yet, it is also possible to identify these dimensions when an assessment is informal and not instrumented or planned, as is often the case with formative assessments. A teacher responding to a question from a learner may use these same dimensions but in a quasi-simultaneous manner. Planning would therefore be minimal, even if competent and experienced teachers would tell us that the questions are often predictable. Teachers would collect data from the answers while judging the value of the question. They would exercise their expert judgment and make a decision (respond with another question, connect the answer with a previously taught concept, etc), and, finally, communicate it to the learner. We could compare these 2 methods to type 1 and type 2 clinical reasoning.7 Type 1 reasoning (rapid and intuitive) relates to the formative context, while type 2 reasoning (logic and analysis) relates to the summative context.
Accept that assessments will never be completely objective
An assessment, whether formative or summative, is first and foremost what we saw in the preceding point: a judgment. By definition, a judgment inherently possesses a subjective component. One might incorrectly think that multiple-choice questions are objective. However, making most of the questions about a small fraction of the material taught is enough to render the examination biased and not reflective of learners’ true progress. The grading process is the only aspect of multiple-choice questions that is objective; the questions are not. In other contexts where observation is required, as in objective structured clinical examinations, several researchers have become interested in conditions that legitimately reduce subjectivity.8 We must accept the fact that assessments are an imperfect human activity. That said, we must also work tirelessly to eliminate obstacles potentially affecting the validity of assessments.
To measure or to observe, that is the question
While assessing learning, we can rely on 2 types of information: measurement and observation. Measurement most frequently refers to marks obtained, for example, on multiple-choice questions. The number of “correct responses” serves as a “measurement”: the higher the number of correct responses, the better the learning. This correlation is not always true, but it is the essence of measuring learning. In preclerkships, learning progress is generally measured. During clerkships and residencies, observations are more heavily relied upon, which requires recording tools such as field notes and criteria grids based on the CanMEDS framework.9 The nature of the information collected, therefore, depends on its context. It is advisable to identify attributes of this context that allow one to develop or rely upon tools adapted to a specific situation.
Conclusion
It is now well documented that formative assessment has a positive impact on learning.2-4 Learning assessment training should be a part of professional development for physicians working in pedagogy.
Teaching tips
▸ The learning assessment is an approach that aims to produce the best possible decisions, both in pedagogic and administrative contexts.
▸ Learning assessment is a discipline in and of itself, with a considerable body of knowledge both on medical education as well as other pedagogic disciplines.
▸ Training on learning assessment should be a part of professional development for physicians working in pedagogy.
Teaching Moment is a quarterly series in Canadian Family Physician, coordinated by the Section of Teachers of the College of Family Physicians of Canada. The focus is on practical topics for all teachers in family medicine, with an emphasis on evidence and best practice. Please send any ideas, requests, or submissions to Dr Viola Antao, Teaching Moment Coordinator, at viola.antao@utoronto.ca.
Footnotes
Competing interests
None declared
La version en français de cet article se trouve à https://www.cfp.ca dans la table des matières du numéro de janvier 2023 à la page 67.
References
- 1.Desrosiers É. Mastering teaching as a resident. Easier done than said. Can Fam Physician 2015;61:e344-6 (Eng), 642-4 (Fr). [PMC free article] [PubMed] [Google Scholar]
- 2.Mitra NK, Barua A.. Effect of online formative assessment on summative performance in integrated musculoskeletal system module. BMC Med Educ 2015;15:29. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Hattie J. Visible learning for teachers. Maximizing impact on learning. New York, NY: Routledge; 2012. [Google Scholar]
- 4.Black P, Wiliam D.. Inside the black box: raising standards through classroom assessment. Phi Delta Kappan 2010;92(1):81-90. [Google Scholar]
- 5.Laurier MD, Tousignant R, Morissette D.. Les principes de la mesure et de l’évaluation des apprentissages. 3rd ed. Montréal, QC: Gaëtan Morin; 2005. [Google Scholar]
- 6.Brousseau G. Fondements et méthodes de la didactique des mathématiques. Recherches en Didactique des Mathématiques 1986;7(2):33-115. [Google Scholar]
- 7.Kahneman D. Thinking, fast and slow. New York, NY: Farrar, Straus and Giroux; 2011. [Google Scholar]
- 8.Chong L, Taylor S, Haywood M, Adelstein BA, Shulruf B.. The sights and insights of examiners in objective structured clinical examinations. J Educ Eval Health Prof 2017;14:34. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.CanMEDS: better standards, better physicians, better care. Ottawa, ON: Royal College of Physicians and Surgeons of Canada; 2015. Available from: https://www.royalcollege.ca/rcsite/canmeds/canmeds-framework-e. Accessed 13 Oct 2022. [Google Scholar]