Skip to main content
Perspectives on Medical Education logoLink to Perspectives on Medical Education
. 2024 Sep 11;13(1):452–459. doi: 10.5334/pme.1150

Validity in the Next Era of Assessment: Consequences, Social Impact, and Equity

Benjamin Kinnear 1,**, Christina St-Onge 2,**, Daniel J Schumacher 3, Mélanie Marceau 4, Thirusha Naidu 5,6
PMCID: PMC11396166  PMID: 39280703

Abstract

Validity has long held a venerated place in education, leading some authors to refer to it as the “sine qua non” or “cardinal virtue” of assessment. And yet, validity has not held a fixed meaning; rather it has shifted in its definition and scope over time. In this Eye Opener, the authors explore if and how current conceptualizations of validity fit a next era of assessment that prioritizes patient care and learner equity. They posit that health profession education’s conceptualization of validity will change in three related but distinct ways. First, consequences of assessment decisions will play a central role in validity arguments. Second, validity evidence regarding impacts of assessment on patients and society will be prioritized. Third, equity will be seen as part of validity rather than an unrelated concept. The authors argue that health professions education has the agency to change its ideology around validity, and to align with values that predominate the next era of assessment such as high-quality care and equity for learners and patients.


Validity has long held a venerated place in education, leading some authors to refer to it as the “sine qua non” or “cardinal virtue” of assessment [1,2]. And yet, validity has not held a fixed meaning; rather it has shifted in its definition and scope over time. How will validity change as health professions education (HPE) assessment evolves? In this Eye Opener, we explore if and how current conceptualizations of validity fit the values in a next era of assessment that focuses on ensuring high-quality care for patients. Specifically, we explore what might be required for validity to support a world in which assessment is more socially accountable and equity-focused.

A brief overview of validity and some contemporary conceptualizations

Validity conceptualizations in HPE have evolved over time. In 2017, St-Onge et al. [3] made explicit three different, co-existing conceptualizations of validity in the HPE literature: validity as a test characteristic, validity as an argument-based evidentiary chain, and validity as a social imperative (a conceptualization still nascent in HPE).

The first conceptualization, validity as a test characteristic, is strongly tied to measurement models, namely Classical Test Theory, Generalizability Theory and Item-Response Theory [4,5]. These theories and models aim to quantify measurement error and infer individuals’ “true” scores [4,6,7]. Reliability and validity are significantly intertwined, with the pursuit of a true score (Classical Test Theory), or a generalizability coefficient (reliability of score given a specified universe of generalization) [4]. Additional pursuit of score precision can be seen in Item Response Theory, which focuses on individual item-level difficulty [5,6]. In this view of validity, quantitative evidence to support an assessment score’s reliability, generalizability, or precision is highly valued. Validity is a characteristic attributed to a test, indicating that “it measured what it intended to measure” [8,9]. This conceptualization of validity still exists in HPE, most often with regard to sellable assessment products.

The second conceptualization, validity as an argument-based evidentiary chain, focuses on documenting the appropriateness of the interpretations and decisions made based on assessment data [3,10]. Two argument-based approaches have been predominantly imported into HPE, Messick’s unified theory of validity [11] and Kane’s approach to validation [12,13]. Authors that imported these approaches into HPE translated abstract validity conceptualizations into more concrete validation practices (e.g., Cook and Hatala [14], Kinnear et al. [15]). With the multi-faceted and complex programs of assessment that are increasingly found in competency-based education (CBE) [16], argument-based approaches allow for multiple, different types of evidence to be developed and integrated into fit-for-purpose arguments about the validity of assessment decisions. Validity as a social imperative (the third conceptualization) has grown out of argument-based approaches and, as we detail below, aligns well with the next era of assessment.

The next era of validity in HPE assessment

The next era of validity will be shaped by broader forces and trends in HPE assessment. As a result, we believe validity will change in three related but distinct ways. First, HPE has already integrated the consequences of assessment decisions into validity conceptualizations, though consequences remain mostly unaddressed in real-world validation work. In the next era, consequences of assessment decisions will play a central role in validity. Second, the proliferation of CBE has foregrounded assessment’s role in social accountability. In the future, validity arguments will be more directly connected to impacts on patients and society. Third, equity considerations have become central in many HPE spaces. Similarly, equity will become central to future validity arguments. While consequences of assessment, social accountability, and equity are not novel concepts in HPE assessment, they are not central to most work on validity. Below we expand on how each will play a central role in the next era of validity.

Focus on consequences of assessment decisions

Messick’s unified theory of validity brought with it the concept of different sources of evidence that can be sought to support the validity of assessment decisions [11]. One such source of evidence was the social consequence of test uses, which Messick called “consequential validity” evidence [17], now sometimes called “consequential evidence” [18]. Cook and Lineberry explored consequential evidence in HPE, describing it as “the impact, beneficial or harmful and intended or unintended, of assessment.” [18]. Consequences include impacts on learners, educators, programs, patients, and other systems and people. The concept of consequential validity evidence has gained acceptance in HPE, regularly appearing in HPE manuscripts describing the concept of validity and the process of validation [1,19,20,21,22,23,24]. Even the oft-cited Standards for Educational and Psychological Testing includes consequential evidence as important for validity arguments [10].

Cook et al. argue that “evidence of consequences is ultimately the most important source of validity evidence” [18]. The authors take a teleological stance, drawing an analogy with clinical diagnostic tests. Regardless of a diagnostic test’s sensitivity or specificity, its ultimate value will depend on consequences to patients, hospitals, and society. Similarly, while all sources of validity evidence have value, consequential evidence should be central to any validity argument. Despite this, consequential evidence is one of the least reported types of validity evidence. Across three systematic reviews on HPE validation work [25,26,27], consequential evidence was reported in only 5–20% of studies [18]. While one cannot say for certain why consequential validity evidence is relatively rare in HPE, contributing reasons likely include challenging study designs, limited resources for validation work, and vestigial preferences for more psychometric data such as reliability (i.e. internal structure) or criterion (i.e. relationship to other variables) evidence.

By making consequential evidence part of (or central to) validity arguments, we expand validity’s reach. Validation becomes more than ‘demonstrating that you are measuring what you think you are measuring’ to also include downstream effects of assessment. The next era of validity should include more widespread integration of consequential evidence into validity arguments. In the following sections, we explore how studying social impact and equity in assessment can provide meaningful consequential evidence.

Evidence of the social impact of assessments

As noted above, consequential validity evidence includes impacts of assessment decisions on patients and society, and hence represents a form of social accountability. CBE, the predominant training philosophy in HPE in many countries, is rooted in social accountability [28,29]. Marceau et al. recently made explicit the concept of validity as a “social imperative” [30,31], in which validation is a mechanism to ensure that assessment decisions are linked to societal impacts. This view brings a deontological lens to consequential validity evidence by emphasizing HPE’s social contract with the patients it ultimately serves. To that end, the next era of validity will require evidence that assessment ensures trainees and graduates of HPE programs are providing high-quality care.

While the connection between HPE and patient outcomes is complex and non-linear [32], promising approaches are emerging to develop such consequential validity evidence. Clinical care measures that are seen as sufficiently attributable to individual trainees are being developed in multiple medical specialties [33,34,35,36]. Improving technology, such as haptics and artificial intelligence, could provide real-time assessment of procedural, communication, and teamwork skills [37]. Better understanding of interdependence of competence could unlock new ways of assessing team-based care outcomes [38,39]. All of these approaches can be integrated into current programs of assessment to bolster social accountability by connecting education to patient care.

Centering equity in assessment validity

By recognizing consequential evidence as essential to validity arguments, we also make equity central to validity. We define equity in assessment as the opportunity to demonstrate and develop one’s knowledge, skills and abilities without negative influence by “structural or interpersonal bias related to personal or social characteristics of learners or assessors.” [40]. Equity goes beyond impartiality and includes efforts to ensure that each learner is afforded the resources and opportunities that they need to succeed, acknowledging that individuals need different types and levels of support and face different societal and system biases [41,42,43,44].

Equity is certainly not new to assessment, with scholars and advocates noting the many biases and injustices that have plagued HPE assessment for years [45,46,47,48,49]. Taking a sociocultural view, assessment has played a key role in creating and maintaining hegemony via control of patronage and access to educational and professional opportunities [50]. Performance on any particular assessment favors the dominant social order which influenced an instrument’s creation, while establishing what is accepted as truth and knowledge [50]. Viewed this way, assessment reinforces power structures while normalizing judgment. Thus, attending to equity is critical to promote fairness and justice for everyone impacted by assessment, particularly those who have been marginalized in a society.

Contemporary advocacy efforts such as #MeToo; Black Lives Matter; and advancing LGBTQ+, Feminist and Indigenous rights have brought equity to the fore of many discourses in HPE. Inequitable assessment is increasingly recognized as a driver of significant and tangible negative effects on learners that amplify and compound over time [45]. Equitable assessment should include choices of assessments that are inclusive of learners who require accommodations [41,51]. Current assessment accommodations often require learners to come forward, self-identify, and justify their requests [52]. However, education systems are rarely designed to help learners feel comfortable enough to do this [41,44]. Inequitable assessment also stands to harm patients by reducing the diversity of healthcare professionals that are available to serve diverse patient populations [53,54,55]. The next era of assessment brings a growing urgency to foreground equity in assessment [41,44,56,57], and validation practices should align with such goals.

Evidence of equity in assessment can be sought by examining the design of assessment tools (i.e. intrinsic equity), the learning environment (i.e. contextual equity), or the uses of assessment data to create equitable opportunities (i.e. instrumental equity) [46]. Onumah et al provide an example of how assessment systems can be designed with all three facets of equity in mind [58]. Equity also means programs should seek to understand how colonialism, racism, and Global North Euro-American principles have shaped HPE’s ideology and propagated inequities [47]. Including equity in validity arguments means that if assessment decisions are shown to worsen inequity for learners or patients, then we deem those decisions not valid.

Notably, we are not implying that equity is secondary or subordinate to validity, nor that equity is only important if examined through the lens of validity. We also do not believe that all of the richness, complexity, and nuance of equity initiatives can be captured within a validity argument. However, equity has long been treated as an afterthought in HPE assessment. By making equity a central part of validity arguments, it too becomes a ‘cardinal virtue’ of assessment. Therefore, in the next era of assessment, equity can function much like the concept of reliability – standing as a distinct concept while also being an integral part of validity arguments.

Broad vs narrow conceptualizations of validity: choosing our ideology

HPE is not a monolithic group, and we do not expect that everyone will agree with our call to foreground consequences, social impact, and equity in assessment validity. Some validity scholars disagree with the assertion that consequential evidence should be part of validity at all, instead advocating that validity should focus only on construct representation, not the downstream impacts of assessment [59,60,61,62,63,64]. We are not implying that those scholars do not care about the consequences of assessment such as societal impact or equity. Rather, they see consequences as being different from validity, to be considered separately under categories such as “utility” or “acceptability”. The varying conceptualizations of validity reflect the many disciplines that comprise HPE (e.g. Psychology, Sociology, Measurement, Education) [65], and this diversity of viewpoints represents a strength for our field. Which view will predominate in the next era of assessment?

The wonderful news is that HPE has agency in such a choice. As Varpio points out, “Fortunately, ideology is maintained by our decisions and actions; therefore, we can change our decisions and thereby modify the ideology to work for us, not against us.” [66]. As the next era unfolds, we can align our conceptualizations of validity with the values underpinning our assessment work. As noted at the outset of the article, validity holds a long and tenured position as being the “sine qua non” of assessment. We believe consequences, social impact and equity deserve the same status in the next era or assessment.

A significant challenge will involve anticipating and identifying all relevant consequences of given assessments. When considering the equity consequences of assessment, where do we begin and end? Equity initiatives often range from efforts to improve equitable access to medical education, to ensuring diverse individuals have proper resources to succeed in training programs, to monitoring for negative downstream consequences of assessment decisions. Real-world decisions will be needed to determine where to focus energy and resources in developing validity arguments with a seemingly infinite amount of evidence to be collected. To navigate these discussions, a socio-constructivist approach to validity could be embraced to co-construct the argument to support the defensibility of these assessment choices. Or perhaps critical theory approaches can help ensure that we are scrutinizing who gets to decide what is equitable, whose voice is being valued, and why.

The next era of validity will require adopting a more inclusive perspective. Most, if not all, current authors and leaders in validity are from Europe and North America [67,68]. Thus, we must be mindful before transposing these validity conceptualizations and practices in other contexts and regions. Anecdotal evidence suggests that Global North customs and practices may not always transpose well in the Global South context. This should be further investigated. We can learn from the current and ongoing challenges of applying Global North conceptualizations in the Global South to challenge our assumptions about validity and validation and inform future development. We should also be open to the idea of validity expanding even further to include not just equity, but ideas such as social responsiveness and awareness. We have agency to determine where the boundaries of validity lie, how they can and should change over time, and which approaches best serve our learners and patients.

Conclusion

The values of consequences, social accountability and equity will significantly influence assessment and validity in their next era. These values will undoubtedly challenge the current approach to validation and may require some to reconsider what falls under the purview of validity. A broader conceptualization of validity and validation that incorporates equity concepts in the purpose, design, and use of assessment data could contribute to assessments that are not just technically and psychometrically sound, but also socially accountable and equitable for learners and patients. Like so many required changes in assessment, changing our conceptualization of validity requires a shift in ideology. The next era in assessment has the potential to catalyze novel ways to develop, share, and evaluate validity arguments with impacts on our patients and learners at the center of what we do.

Competing Interests

The authors have no competing interests to declare.

References

  • 1.Downing SM. Validity: on the meaningful interpretation of assessment data. Med Educ. 2003; 37(9): 830–7. DOI: 10.1046/j.1365-2923.2003.01594.x [DOI] [PubMed] [Google Scholar]
  • 2.Mislevy RJ, Steinberg LS, Almond RG. Focus article: On the structure of educational assessments. Measurement: Interdisciplinary research and perspectives. 2003; 1(1): 3–62. DOI: 10.1207/S15366359MEA0101_02 [DOI] [Google Scholar]
  • 3.St-Onge C, Young M, Eva KW, Hodges B. Validity: one word with a plurality of meanings. Adv Health Sci Educ Theory Pract. 2017; 22(4): 853–67. DOI: 10.1007/s10459-016-9716-3 [DOI] [PubMed] [Google Scholar]
  • 4.Cronbach LJ, Rajaratnam N, Gleser GC. Theory of generalizability: A liberalization of reliability theory. Br J Stat Psychol. 1963; 16(2): 137–63. DOI: 10.1111/j.2044-8317.1963.tb00206.x [DOI] [Google Scholar]
  • 5.Schuwirth LW, van der Vleuten CP. General overview of the theories used in assessment: AMEE Guide No. 57. Med Teach. 2011; 33(10): 783–97. DOI: 10.3109/0142159X.2011.611022 [DOI] [PubMed] [Google Scholar]
  • 6.Hambleton RK, Swaminathan H. Item response theory: Principles and applications. Springer Science & Business Media; 2013. [Google Scholar]
  • 7.Embretson SE, Reise SP. Item response theory. Psychology Press; 2013. DOI: 10.4324/9781410605269 [DOI] [Google Scholar]
  • 8.Smith HL, Wright WW. Tests and measurements. Silver, Burdett; 1928. [Google Scholar]
  • 9.Garrett HE. Statistics in psychology and education. Longmans, Green and Company; 1926. [Google Scholar]
  • 10.Standards for educational and psychological testing. Amer Educational Research Assn; 2014. [Google Scholar]
  • 11.Messick S. Validity. In Linn R. (Ed.) Educational measurement (13–103). New York: Macmillan Publishing; 1989. [Google Scholar]
  • 12.Kane MT. An argument-based approach to validity. Psychol Bull. 1992; 112(3): 527. DOI: 10.1037/0033-2909.112.3.527 [DOI] [Google Scholar]
  • 13.Kane MT. Validating the interpretations and uses of test scores. J Educ Meas. 2013; 50(1): 1–73. DOI: 10.1111/jedm.12000 [DOI] [Google Scholar]
  • 14.Cook DA, Hatala R. Validation of educational assessments: a primer for simulation and beyond. Adv Simul (Lond). 2016; 1: 31. DOI: 10.1186/s41077-016-0033-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Kinnear B, Kelleher M, May B, Sall D, Schauer DP, Schumacher DJ, et al. Constructing a validity map for a workplace-based assessment system: Cross-walking Messick and Kane. Acad Med. 2021; 96(7S): S64–S9. DOI: 10.1097/ACM.0000000000004112 [DOI] [PubMed] [Google Scholar]
  • 16.Schuwirth LW, van der Vleuten CP. Programmatic assessment and Kane’s validity perspective. Med Educ. 2012; 46(1): 38–48. DOI: 10.1111/j.1365-2923.2011.04098.x [DOI] [PubMed] [Google Scholar]
  • 17.Messick S. Validity of test interpretation and use. Research Report for the Educational Testing Service; 1990. DOI: 10.1002/j.2333-8504.1990.tb01343.x [DOI] [Google Scholar]
  • 18.Cook DA, Lineberry M. Consequences Validity Evidence: Evaluating the Impact of Educational Assessments. Acad Med. 2016; 91(6): 785–95. DOI: 10.1097/ACM.0000000000001114 [DOI] [PubMed] [Google Scholar]
  • 19.Sullivan GM. A primer on the validity of assessment instruments. Sullivan GM. A primer on the validity of assessment instruments. J Grad Med Educ. 2011. Jun 1; 3(2): 119–20. DOI: 10.4300/JGME-D-11-00075.1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Peeters MJ, Martin BA. Validation of learning assessments: a primer. Curr Pharm Teach Learn. 2017; 9(5): 925–33. DOI: 10.1016/j.cptl.2017.06.001 [DOI] [PubMed] [Google Scholar]
  • 21.Cook DA, Beckman TJ. Current concepts in validity and reliability for psychometric instruments: theory and application. Am J Med. 2006; 119(2): 166 e7–16. DOI: 10.1016/j.amjmed.2005.10.036 [DOI] [PubMed] [Google Scholar]
  • 22.Andreatta PB, Gruppen LD. Conceptualising and classifying validity evidence for simulation. Med Educ. 2009; 43(11): 1028–35. DOI: 10.1111/j.1365-2923.2009.03454.x [DOI] [PubMed] [Google Scholar]
  • 23.Gasmalla HEE, Tahir ME. The validity argument: Addressing the misconceptions. Med Teach. 2021; 43(12): 1453–5. DOI: 10.1080/0142159X.2020.1856802 [DOI] [PubMed] [Google Scholar]
  • 24.Noureldin YA, Lee JY, McDougall EM, Sweet RM. Competency-based training and simulation: making a “valid” argument. J Endour. 2018; 32(2): 84–93. DOI: 10.1089/end.2017.0650 [DOI] [PubMed] [Google Scholar]
  • 25.Cook DA, Zendejas B, Hamstra SJ, Hatala R, Brydges R. What counts as validity evidence? Examples and prevalence in a systematic review of simulation-based assessment. Adv Health Sci Educ Theory Pract. 2014; 19(2): 233–50. DOI: 10.1007/s10459-013-9458-4 [DOI] [PubMed] [Google Scholar]
  • 26.Beckman TJ, Cook DA, Mandrekar JN. What is the validity evidence for assessments of clinical teaching? J Gen Intern Med. 2005; 20: 1159–64. DOI: 10.1111/j.1525-1497.2005.0258.x [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Kogan JR, Holmboe ES, Hauer KE. Tools for direct observation and assessment of clinical skills of medical trainees: a systematic review. JAMA. 2009; 302(12): 1316–26. DOI: 10.1001/jama.2009.1365 [DOI] [PubMed] [Google Scholar]
  • 28.Morcke AM, Dornan T, Eika B. Outcome (competency) based education: an exploration of its origins, theoretical basis, and empirical evidence. Adv Health Sci Educ Theory Pract. 2013; 18(4): 851–63. DOI: 10.1007/s10459-012-9405-9 [DOI] [PubMed] [Google Scholar]
  • 29.Hodge S. The Origins of Competency-Based Training. Aust J Adult Learn. 2007; 47(2): 179–209. [Google Scholar]
  • 30.Marceau M, Gallagher F, Young M, St-Onge C. Validity as a social imperative for assessment in health professions education: a concept analysis. Med Educ. 2018; 52(6): 641–53. DOI: 10.1111/medu.13574 [DOI] [PubMed] [Google Scholar]
  • 31.Marceau M, St-Onge C, Gallagher F, Young M. Validity as a social imperative: users’ and leaders’ perceptions. Can Med Educ J. 2022; 13(3): 22–36. DOI: 10.36834/cmej.73518 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Varpio L, Sherbino J. Demonstrating causality, bestowing honours, and contributing to the arms race: Threats to the sustainability of HPE research. Med Educ. 2024; 58(1): 157–63. DOI: 10.1111/medu.15148 [DOI] [PubMed] [Google Scholar]
  • 33.Schumacher DJ, Martini A, Holmboe E, Carraccio C, van der Vleuten C, Sobolewski B, et al. Initial implementation of resident-sensitive quality measures in the pediatric emergency department: A wide range of performance. Acad Med. 2020; 95(8): 1248–55. DOI: 10.1097/ACM.0000000000003147 [DOI] [PubMed] [Google Scholar]
  • 34.Schumacher DJ, Holmboe ES, van der Vleuten C, Busari JO, Carraccio C. Developing resident-sensitive quality measures: A model from pediatric emergency medicine. Acad Med. 2018; 93(7): 1071–8. DOI: 10.1097/ACM.0000000000002093 [DOI] [PubMed] [Google Scholar]
  • 35.Burk-Rafel J, Sebok-Syer SS, Santen SA, Jiang J, Caretta-Weyer HA, Iturrate E, et al. TRainee Attributable & Automatable Care Evaluations in Real-time (TRACERs): A Scalable Approach for Linking Education to Patient Care. Perspect Med Educ. 2023; 12(1): 149. DOI: 10.5334/pme.1013 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Kalet AL, Gillespie CC, Schwartz MD, Holmboe ES, Ark TK, Jay M, et al. New measures to establish the evidence base for medical education: identifying educationally sensitive patient outcomes. Acad Med. 2010; 85(5): 844–51. DOI: 10.1097/ACM.0b013e3181d734a5 [DOI] [PubMed] [Google Scholar]
  • 37.Kinnear B, Caretta-Weyer H, Lam AC, Tang B, Ginsburg S, Wong BM, et al. Finding Medicine’s Moneyball: How Lessons from Major League Baseball Can Advance Assessment in Precision Education. Acad Med. 2023: 10.1097. [DOI] [PubMed] [Google Scholar]
  • 38.Sebok-Syer SS, Chahine S, Watling CJ, Goldszmidt M, Cristancho S, Lingard L. Considering the interdependence of clinical performance: implications for assessment and entrustment. Med Educ. 2018; 52(9): 970–80. DOI: 10.1111/medu.13588 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Sebok-Syer SS, Lingard L, Panza M, Van Hooren TA, Rassbach CE. Supportive and collaborative interdependence: Distinguishing residents’ contributions within health care teams. Med Educ. 2023 Oct; 57(10): 921–31. DOI: 10.1111/medu.15064 [DOI] [PubMed] [Google Scholar]
  • 40.Rashid MA, Ali SM, Dharanipragada K. Decolonising medical education regulation: a global view. BMJ Glob Heal. 2023; 8(6): e011622. DOI: 10.1136/bmjgh-2022-011622 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Ajjawi R, Tai J, Boud D, Jorre de St Jorre T. Assessment for inclusion in higher education. Promoting equity and social justice in assessment: Taylor & Francis; 2023. DOI: 10.4324/9781003293101 [DOI] [Google Scholar]
  • 42.Aquario D. Through the lens of justice. A systematic review on equity and fairness in learning assessment. Educ Sci Soc. 2021(2021/2). DOI: 10.3280/ess2-2021oa12405 [DOI] [Google Scholar]
  • 43.Hanesworth P, Bracken S, Elkington S. A typology for a social justice approach to assessment: learning from universal design and culturally sustaining pedagogy. Teach High Educ. 2019; 24(1): 98–114. DOI: 10.1080/13562517.2018.1465405 [DOI] [Google Scholar]
  • 44.Tai J, Ajjawi R, Boud D, de St Jorre TJ. Promoting Equity and Social Justice through Assessment for Inclusion. Assessment for Inclusion in Higher Education. Routledge; 2022. p. 9–18. DOI: 10.4324/9781003293101-3 [DOI] [Google Scholar]
  • 45.Teherani A, Hauer KE, Fernandez A, King TE Jr., Lucey C. How small differences in assessed clinical performance amplify to large differences in grades and awards: a cascade with serious consequences for students underrepresented in medicine. Acad Med. 2018; 93(9): 1286–92. DOI: 10.1097/ACM.0000000000002323 [DOI] [PubMed] [Google Scholar]
  • 46.Lucey CR, Hauer KE, Boatright D, Fernandez A. Medical education’s wicked problem: achieving equity in assessment for medical learners. Acad Med. 2020; 95(12S): S98–S108. DOI: 10.1097/ACM.0000000000003717 [DOI] [PubMed] [Google Scholar]
  • 47.Anderson N, Nguyen M, Marcotte K, Ramos M, Gruppen LD, Boatright D. The long shadow: a historical perspective on racism in medical education. Acad Med. 2023; 98(8S): S28–S36. DOI: 10.1097/ACM.0000000000005253 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Anderson H, Abdulla L, Balmer D, Govaerts M, Busari J. Inequity is woven into the fabric: a discourse analysis of assessment in pediatric residency training. Adv Health Sci Educ Theory Pract. 2024. Mar; 29(1): 199–216. DOI: 10.1007/s10459-023-10260-9 [DOI] [PubMed] [Google Scholar]
  • 49.Gipps C. What do we mean by equity in relation to assessment? Assess Educ. 1995; 2(3): 271–81. DOI: 10.1080/0969595950020303 [DOI] [Google Scholar]
  • 50.Gipps C. Chapter 10: Socio-cultural aspects of assessment. Rev Res Educ. 1999; 24(1): 355–92. DOI: 10.3102/0091732X024001355 [DOI] [Google Scholar]
  • 51.McArthur J. Assessment for social justice: The role of assessment in achieving social justice. Assess Eval High Educ. 2016; 41(7): 967–81. DOI: 10.1080/02602938.2015.1053429 [DOI] [Google Scholar]
  • 52.Kinnear B, Weber DE, Schumacher DJ, Edje L, Warm EJ, Anderson HL. Reconstructing Neurath’s Ship: A Case Study in Reevaluating Equity in a Program of Assessment. Acad Med. 2023. Aug 1; 98(8S): S50–6. DOI: 10.1097/ACM.0000000000005249 [DOI] [PubMed] [Google Scholar]
  • 53.Snyder JE, Upton RD, Hassett TC, Lee H, Nouri Z, Dill M. Black representation in the primary care physician workforce and its association with population life expectancy and mortality rates in the US. JAMA Net Open. 2023; 6(4): e236687–e. DOI: 10.1001/jamanetworkopen.2023.6687 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54.Cooper LA, Roter DL, Johnson RL, Ford DE, Steinwachs DM, Powe NR. Patient-centered communication, ratings of care, and concordance of patient and physician race. Ann Intern Med. 2003; 139(11): 907–15. DOI: 10.7326/0003-4819-139-11-200312020-00009 [DOI] [PubMed] [Google Scholar]
  • 55.Saha S, Komaromy M, Koepsell TD, Bindman AB. Patient-physician racial concordance and the perceived quality and use of health care. Arch Intern Med. 1999; 159(9): 997–1004. DOI: 10.1001/archinte.159.9.997 [DOI] [PubMed] [Google Scholar]
  • 56.Holmboe ES, Osman NY, Murphy CM, Kogan JR. The urgency of now: Rethinking and improving assessment practices in medical education programs. Acad Med. 2023; 98(8S): S37–S49. DOI: 10.1097/ACM.0000000000005251 [DOI] [PubMed] [Google Scholar]
  • 57.Larson T. Macy Josiah Jr.. Foundation Conference on Ensuring Fairness in Medical Education Assessment: Conference Recommendations Report. Acad Med. 2023; 98(8S): S3–S15. DOI: 10.1097/ACM.0000000000005243 [DOI] [PubMed] [Google Scholar]
  • 58.Onumah CM, Pincavage AT, Lai CJ, Levine DL, Ismail NJ, Alexandraki I, et al. Strategies for advancing equity in frontline clinical assessment. Acad Med. 2023; 98(8S): S57–S63. DOI: 10.1097/ACM.0000000000005246 [DOI] [PubMed] [Google Scholar]
  • 59.Clauser BE, Margolis MJ, Holtman MC, Katsufrakis PJ, Hawkins RE. Validity considerations in the assessment of professionalism. Adv Health Sci Educ Theory Pract. 2012; 17: 165–81. DOI: 10.1007/s10459-010-9219-6 [DOI] [PubMed] [Google Scholar]
  • 60.Borsboom D. Whose consensus is it anyway? Scientific versus legalistic conceptions of validity. Meas Interdiscip Res Perspect. 2012; 10(1–2): 38–41. DOI: 10.1080/15366367.2012.681971 [DOI] [Google Scholar]
  • 61.Borsboom D, Wijsen LD. Frankenstein’s validity monster: The value of keeping politics and science separated. Assess Educ Princ Policy Pract. 2016; 23(2): 281–3. DOI: 10.1080/0969594X.2016.1141750 [DOI] [Google Scholar]
  • 62.Newton PE, Shaw SD. Disagreement over the best way to use the word ‘validity’ and options for reaching consensus. Assess Educ Princ Policy Pract. 2016; 23(2): 178–97. DOI: 10.1080/0969594X.2015.1037241 [DOI] [Google Scholar]
  • 63.Newton PE, Baird J-A. The great validity debate. Taylor & Francis; 2016. DOI: 10.1080/0969594X.2016.1172871 [DOI] [Google Scholar]
  • 64.Newton PE. Questioning the consensus definition of validity. Meas Interdiscip Res Perspect. 2012; 10(1–2): 110–22. DOI: 10.1080/15366367.2012.688456 [DOI] [Google Scholar]
  • 65.Norman G. Fifty years of medical education research: waves of migration. Med Educ. 2011; 45(8): 785–91. DOI: 10.1111/j.1365-2923.2010.03921.x [DOI] [PubMed] [Google Scholar]
  • 66.Varpio L. Dismantling medical education’s incompatible ideology. Med Teach. 2023: 1–6. [DOI] [PubMed] [Google Scholar]
  • 67.Young M, St-Onge C, Xiao J, Lachiver EV, Torabi N. Characterizing the literature on validity and assessment in medical education: a bibliometric study. Perspect Med Educ. 2018; 7(3): 182–91. DOI: 10.1007/S40037-018-0433-X [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 68.Kinnear B, Martini A, Varpio L, Driessen EW, Schumacher DJ. How do validity experts conceptualise argumentation? It’s a rhetorical question. Med Educ; 2024. DOI: 10.1111/medu.15311 [DOI] [PubMed] [Google Scholar]

Articles from Perspectives on Medical Education are provided here courtesy of Ubiquity Press

RESOURCES