Skip to main content
Journal of Graduate Medical Education logoLink to Journal of Graduate Medical Education
. 2017 Apr;9(2):209–214. doi: 10.4300/JGME-D-16-00282.1

Using the Entrustable Professional Activities Framework in the Assessment of Procedural Skills

Debra Pugh 1,, Rodrigo B Cavalcanti 1, Samantha Halman 1, Irene W Y Ma 1, Maria Mylopoulos 1, David Shanks 1, Lynfa Stroud 1
PMCID: PMC5398142  PMID: 28439355

Abstract

Background

The entrustable professional activity (EPA) framework has been identified as a useful approach to assessment in competency-based education. To apply an EPA framework for assessment, essential skills necessary for entrustment to occur must first be identified.

Objective

Using an EPA framework, our study sought to (1) define the essential skills required for entrustment for 7 bedside procedures expected of graduates of Canadian internal medicine (IM) residency programs, and (2) develop rubrics for the assessment of these procedural skills.

Methods

An initial list of essential skills was defined for each procedural EPA by focus groups of experts at 4 academic centers using the nominal group technique. These lists were subsequently vetted by representatives from all Canadian IM training programs through a web-based survey. Consensus (more than 80% agreement) about inclusion of each item was sought using a modified Delphi exercise. Qualitative survey data were analyzed using a framework approach to inform final assessment rubrics for each procedure.

Results

Initial lists of essential skills for procedural EPAs ranged from 10 to 24 items. A total of 111 experts completed the national survey. After 2 iterations, consensus was reached on all items. Following qualitative analysis, final rubrics were created, which included 6 to 10 items per procedure.

Conclusions

These EPA-based assessment rubrics represent a national consensus by Canadian IM clinician educators. They provide a practical guide for the assessment of procedural skills in a competency-based education model, and a robust foundation for future research on their implementation and evaluation.

Introduction

Over the past decade, there has been a paradigm shift in health professions education toward a competency-based education (CBE) model.1,2 One of the challenges in CBE has been the increased need to provide robust assessment of competencies achieved by trainees, especially for workplace, criterion-based assessment.3

A key element of CBE is the expectation that graduates become competent in performing procedures, with certifying boards in Canada and the United States having specific procedural skills requirements.4,5 Still, many trainees do not master these skills during residency and report feeling ill-prepared to perform procedures,68 and this may be partly due to limited opportunities for formal assessment during training.9

Studies of assessment methods for procedural skills have largely focused on technical ability, and have not considered nontechnical skills that are equally critical for successful procedural performance.1013 These skills include the ability to communicate effectively (eg, obtaining informed consent); to collaborate with other professionals; to act in a professional manner; and to exercise clinical reasoning and judgment.14 There have been attempts to assess procedural skills in context,1517 but it remains unclear which nontechnical skills are most important to assess.

The entrustable professional activities (EPAs) framework is 1 approach to assessment in a CBE model.18,19 EPAs are observable, holistic activities that require the integration of knowledge, skills, and attitudes across competency domains,20 and they also ground competencies within a physician's day-to-day work. EPAs differ from competencies in that they refer to the components of physicians' work, while competencies refer to the abilities of physicians.21 Examples of EPAs include managing intravenous fluids, disclosing a medical error,22 or performing a lumbar puncture.

Although there are published reports on how to develop EPAs,2226 there is relatively little guidance on how to use them in assessment. One approach entails use of entrustability-aligned anchor scales,27 which reflect the continuum of abilities ranging from passive observer to being able to perform an activity independently and provide supervision to others. Rating scales that use entrustability anchors produce more reliable scores than conventional anchors.28

To date, there is only a small amount of literature to guide the assessment of procedural skills within an EPA framework. Conceptually, using entrustability-aligned anchors makes sense in the assessment of procedural skills because they rely on an expert's holistic judgment regarding whether or not a trainee can be entrusted to perform a procedure (or components thereof). When compared with traditional assessment methods, which tend to focus on a series of checklist items and/or ratings, an entrustment framework may better reflect the realities of clinical practice.

In other words, a supervisor may not trust the abilities of a trainee even if they have received a high score on an examination of procedural skills (eg, if they missed 1 key step, or if the examination did not assess all the competencies required in performing the procedure). Instead, the supervisor makes an overall judgment about a trainee's ability based on a number of sources of information related to the essential skills required to successfully complete the procedure.29 For EPAs to be used in a program of assessment, the essential skills necessary for entrustment to occur must first be identified. In this study, we aimed to (1) define the essential skills required for entrustment to occur for 7 bedside procedures, and (2) develop an EPA-based assessment tool for procedural skills.

Methods

In this multisite study, we defined EPAs for 7 bedside procedures: central venous catheter insertion, lumbar puncture, peripheral arterial catheter insertion, paracentesis, endotracheal intubation, thoracentesis, and knee arthrocentesis. The process for defining the components (ie, essential skills) of the procedural EPAs was divided into 3 phases.

Phase 1: Defining EPA Components

Using a purposive, maximum variation sampling strategy (aimed at capturing a wide range of perspectives),30 the investigators recruited subject matter experts (SMEs) at each of the 4 participating sites (University of British Columbia, University of Calgary, University of Toronto, and University of Ottawa) to participate in focus groups. The SMEs were clinicians with expertise in performing the procedure of interest, who were involved in teaching and/or assessment of residents' procedural skills. To ensure a sample with maximum variation, SMEs included physicians from a variety of specialties, who practiced in both academic and community settings, and who performed procedures in different contexts (eg, intensive care unit versus ambulatory clinic).

Data Collection—Focus Groups: 

Although the essential skills or components required for procedures may be generic, we opted to define the components for all bedside procedures separately and to explore potential differences. We assembled 7 focus groups (1 for each procedure) and asked them to define the components for procedural-specific EPAs. The focus groups were divided among the participating sites (ie, 1 or 2 focus groups per site).

Participants were asked to generate a list of the essential skills required for entrustment to occur (to allow a resident to perform a procedure without supervision). Essential skills required for each EPA were defined using the nominal group technique.31,32 Three iterations were required to achieve consensus on all items. From this information, a final list of EPA components was compiled.

Data Analysis—Focus Group Field Notes: 

After generating the final list of EPA components, the field notes from each focus group were analyzed using a framework approach,33 with a focus on participants' rationales for accepting or rejecting potential EPA components. This analysis helped to inform the development of the EPA component lists for a national survey.

Phase 2: National Survey

To ensure representation from eventual end users, we invited SMEs representing all Canadian internal medicine (IM) and general IM training programs to complete a web-based survey (SurveyMonkey, FluidSurveys, Ottawa, Ontario, Canada). Using a snowball sampling technique, program directors were contacted and asked to identify 8 to 12 practicing physicians involved in their procedural skills curriculum (development, teaching, or assessment) and 8 to 12 residents with experience in performing the procedure of interest, including IM or general IM residents (postgraduate year 3 and above).

National Survey: 

A survey was developed comprising the final list of EPA components for each procedure developed in phase 1. The survey was designed and revised by all coauthors to capture consensus on the proposed EPAs based on small group sessions. Responses indicating that items “should not be included” were followed by a text box allowing respondents to expand on their rationale for disagreement (see online supplemental material for a copy of the survey). For each EPA component that they believed should be excluded from the final list, they were asked to justify their decision with written comments. Demographic information (staff versus resident and university affiliation) was also captured for reporting purposes. The survey was pilot tested by the authors.

What was known and gap

The entrustable professional activity (EPA) framework may be useful in a competency-based education (CBE) model, but little is known about how to use EPAs in the assessment of procedural skills.

What is new

Procedural assessment rubrics based on EPAs developed through a national consensus of Canadian internal medicine clinician educators.

Limitations

Inherent subjectivity of the Delphi approach and response bias in survey research.

Bottom line

The rubrics provide a practical guide for the assessment of procedural skills within a competency-based education model.

Experts identified by IM program directors were sent a link to the web-based survey. For each EPA component, agreement was defined as at least 80% consensus. A follow-up survey was distributed that included only items for which consensus had not yet been reached.

Analysis: 

Response and consensus percentages rates were calculated for each procedure. Qualitative survey comments were analyzed using a framework approach,33 with a particular focus on exploring participants' rationale for the exclusion of potential EPA components.

Phase 3: EPA Rubric Development

In the development of a final EPA rubric for each procedure, the initial lists of EPA components were modified by the investigators based on comments from participants in the national survey. The final lists of EPA components were then used to create the assessment rubrics.

Ethics approval was obtained from each of the 4 participating sites (the universities of British Columbia, Calgary, Toronto, and Ottawa).

Results

Five groups had 8 participants each, 1 group had 7 participants (thoracentesis), and 1 group had 10 participants (paracentesis). Participants (N = 57) were from the specialties of IM, anesthesiology, emergency medicine, and surgery. Eight participants were residents (fourth and fifth years), 1 was a respiratory therapist, and the remainder were practicing physicians, including 3 practicing in a community setting.

The qualitative analysis resulted in refinements of the lists by either subdividing or collapsing items. In addition, stylistic changes (eg, using a verb at the beginning of each component) were made to help standardize the lists. Following this, the revised lists included a range of 12 (thoracentesis) to 21 (central venous catheter insertion) components for each EPA.

A request to distribute a link to the web-based survey was sent to all Canadian IM (n = 17) and general IM (n = 16) program directors. A total of 111 physicians responded (50 practicing physicians and 61 residents). Responses were received from 14 different schools, with respondents from each university ranging from 1 to 18, with a mean of 8.

Consensus was reached on 101 of 102 items (99%) with the first iteration of the survey. For 20 items, there was 100% consensus. By the second iteration (54 respondents), consensus was reached on the 1 outstanding item. In all instances, the consensus was to include the item.

Following this, the qualitative analysis of comments was used to inform any further changes to the final lists. Themes (table) were identified independently by 2 of the coinvestigators (D.P. and M.M.). Using this analysis, final lists of EPA components were modified to produce a rubric to guide the assessment of procedural competence in IM residents. Modifications included adding headings for competencies34 (ie, medical expert, communicator, and professional) and grouping items that were thought to be linked (eg, performing the technical aspects and maintaining sterility).

Table.

Participants' Rationales for Excluding Entrustable Professional Activity Components

graphic file with name i1949-8357-9-2-209-t01.jpg

One item (“knowing how to rewire a line”) was removed from the central line list because 8 respondents identified it as being below the standard of care. For the item “ensuring the availability of an assistant” for an arterial line insertion, 7 respondents noted that an assistant was not always necessary, and the item was revised to include “if required.” Two items (“demonstrating concern for patient comfort and well-being” and “demonstrating an ability to communicate and behave professionally”) were removed based on comments from participants, noting these skills (ie, professionalism, empathy) would apply to all patient interactions.

Finally, we standardized wording across procedures for common items (eg, “obtaining informed consent”) and merged related items for knee arthrocentesis into a single item of “knowing indications/contraindications.”

Entrustability-aligned anchors (ranging from “no task execution” to “supervision may be provided to junior learners”) were then added to the rubric, as adapted from previously published tools.23,25 Final rubrics included 6 to 10 items per procedure (see online supplemental material for a copy of the rubrics).

Discussion

Our study used an EPA framework to develop rubrics to assess IM residents' performance of 7 bedside procedures. These procedural EPAs represent, to our knowledge, the first consensus-driven framework that conceptualizes and delineates the different skills necessary for procedural competence (ie, beyond the technical checklists currently in use). The solicitation of feedback from stakeholders was part of a broader knowledge translation approach that emphasizes the inclusion of knowledge users. The acceptability of an assessment tool is an important factor in determining its use.35 By engaging eventual end users, the final rubrics are more likely to be accepted and incorporated as a tool across programs.

This study used the nominal group technique and a modified Delphi procedure to define the essential skills required for entrustment of residents performing procedures. All items included in the initial lists of EPA components developed by local focus groups were rated as being important by the majority of respondents in a national survey. This adds value, because insights from a diverse group ultimately influenced rubric development and provided an important source of validity evidence (content evidence)36 that contributes to the acceptability of the final product.

An important question when developing EPAs relates to how detailed they should be. EPAs are meant to be relatively broad; therefore, it is necessary to provide some guidance about what, exactly, the EPA encompasses. EPAs can be disaggregated into a list of essential skills that can be mapped to competencies.24,29 However, as EPAs are deconstructed, the risk is that they may become too checklist-driven. In this study, the risk was mitigated by grouping the EPA components into categories so that the number of items in each final EPA rubric was kept to a manageable number.

Our rubrics for assessing procedural skills are to be used as a tool to guide entrustment decisions as residents progress through training.24 The use of entrustability scales allows the level of supervision required to be made explicit, thus helping to ground entrustment decisions. Although they are meant to be used when skills are directly observed, they can be used, or adapted for use, in multiple settings, including when trainees perform procedures on part task models, hybrid settings, or real patients. Importantly, the rubrics allow learners to be assessed along a continuum as they grow in competence. This allows for evaluation of the progression of competency for a particular trainee and facilitates teaching, assessment, and eventual remediation.

Our study serves as a template for the development of other EPA-based competency assessment tools. Although this study focused on the development of EPA-based tools for the procedures performed by internists, the methods used could be applied to other procedural skills, and the developed tools could be adapted to procedures performed by other specialties.

Limitations of this study include the inherent subjectivity of the modified Delphi approach. In addition, the survey data should be interpreted with caution as it is possible that participants' responses were influenced by the lists provided (ie, confirmation bias).

Conclusion

As accreditation bodies in medical education move toward a CBE model, new assessment frameworks are needed to ensure progression and attainment of competency in different domains. The EPA-based assessment rubrics developed in this study represent a national consensus by Canadian clinician educators in IM. The rubrics provide a practical guide for the assessment of procedural skills within a CBE model and a robust foundation for future studies of their implementation and evaluation.

Supplementary Material

References

  • 1. Iobst WF, Sherbino J, ten Cate O, et al. Competency-based medical education in postgraduate medical education. Med Teach. 2010; 32 8: 651– 656. [DOI] [PubMed] [Google Scholar]
  • 2. Lurie SJ. . History and practice of competency-based assessment. Med Educ. 2012; 46 1: 49– 57. [DOI] [PubMed] [Google Scholar]
  • 3. Holmboe ES, Sherbino J, Long DM, et al. The role of assessment in competency-based medical education. Med Teach. 2010; 32 8: 676– 682. [DOI] [PubMed] [Google Scholar]
  • 4. Royal College of Physicians and Surgeons of Canada. Objectives of training in internal medicine, Version 1.0; 2011. http://www.royalcollege.ca/cs/groups/public/documents/document/y2vk/mdaw/∼edisp/tztest3rcpsced000910.pdf. Accessed December 27, 2016. [Google Scholar]
  • 5. American Board of Internal Medicine. Procedures required for internal medicine. http://www.abim.org/certification/policies/imss/im.aspx#procedures. Accessed December 27, 2016. [Google Scholar]
  • 6. Card SE, Snell L, O'Brien B. . Are Canadian general internal medicine training program graduates well prepared for their future careers? BMC Med Educ. 2006; 6: 56. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7. Huang GC, Smith CC, Gordon CE, et al. Beyond the comfort zone: residents assess their comfort performing inpatient medical procedures. Am J Med. 2006; 119 1: 71.e17– e24. [DOI] [PubMed] [Google Scholar]
  • 8. Mourad M, Kohlwes J, Maselli J, et al. Supervising the supervisors—procedural training and supervision in internal medicine residency. J Gen Intern Med. 2010; 25 4: 351– 356. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9. Pugh D, Touchie C, Code C, Humphrey-Murto S. . Teaching and testing procedural skills: survey of Canadian internal medicine program directors and residents [abstract 76]. In: Proceedings from the International Conference on Residency Education; Ottawa ON, Canada; 2010; 4 suppl 3: 29. [Google Scholar]
  • 10. Barsuk JH, McGaghie WC, Cohen ER, et al. Use of simulation-based mastery learning to improve the quality of central venous catheter placement in a medical intensive care unit. J Hosp Med. 2009; 4 7: 397– 403. [DOI] [PubMed] [Google Scholar]
  • 11. Wayne DB, Barsuk JH, O'Leary KJ, et al. Mastery learning of thoracentesis skills by internal medicine residents using simulation technology and deliberate practice. J Hosp Med. 2008; 3 1: 48– 54. [DOI] [PubMed] [Google Scholar]
  • 12. Ma IW, Zalunardo N, Pachev G, et al. Comparing the use of global rating scale with checklists for the assessment of central venous catheterization skills using simulation. Adv Health Sci Educ Theory Pract. 2012; 17 4: 457– 470. [DOI] [PubMed] [Google Scholar]
  • 13. Ma IW, Sharma N, Brindle ME, et al. Measuring competence in central venous catheterization: a systematic review. Springerplus. 2014; 3: 33. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14. Pugh D, Hamstra SJ, Wood TJ, et al. A procedural skills OSCE: assessing technical and non-technical skills of internal medicine residents. Adv Health Sci Educ Theory Pract. 2015; 20 10: 85– 100. [DOI] [PubMed] [Google Scholar]
  • 15. Nestel D, Kneebone R. . Perspective: authentic patient perspectives in simulations for procedural and surgical skills. Acad Med. 2010; 85 5: 889– 893. [DOI] [PubMed] [Google Scholar]
  • 16. Kneebone R, Nestel D, Yadollahi F, et al. Assessing procedural skills in context: exploring the feasibility of an Integrated Procedural Performance Instrument (IPPI). Med Educ. 2006; 40 11: 1105– 1114. [DOI] [PubMed] [Google Scholar]
  • 17. Stroud L, Cavalcanti RB. . Hybrid simulation for knee arthrocentesis: improving fidelity in procedures training. J Gen Intern Med. 2013; 28 5: 723– 727. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18. Royal College of Physicians and Surgeons of Canada. About milestones and entrustable professional activities. http://www.royalcollege.ca/rcsite/cbd/milestones-entrustable-professional-acts-epas-e. Accessed December 27, 2016. [Google Scholar]
  • 19. ten Cate O. . AM last page: what entrustable professional activities add to a competency-based curriculum. Acad Med. 2014; 89 4: 691. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20. ten Cate O. . Trust, competence, and the supervisor's role in postgraduate training. BMJ. 2006; 333 7571: 748– 751. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21. Hauer KE, Soni K, Cornett P, et al. Developing entrustable professional activities as the basis for assessment of competence in an internal medicine residency: a feasibility study. J Gen Intern Med. 2013; 28 8: 1110– 1114. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22. Touchie C, De Champlain A, Pugh D, et al. Supervising incoming first-year residents: faculty expectations versus residents' experiences. Med Educ. 2014; 48 9: 921– 929. [DOI] [PubMed] [Google Scholar]
  • 23. ten Cate O, Hart D, Ankel F, et al. Entrustment decision making in clinical training. Acad Med. 2016; 91 2: 1– 8. [DOI] [PubMed] [Google Scholar]
  • 24. ten Cate O, Chen HC, Hoff RG, et al. Curriculum development for the workplace using entrustable professional activities (EPAs): AMEE Guide No. 99. Med Teach. 2015; 37 11: 983– 1002. [DOI] [PubMed] [Google Scholar]
  • 25. Aylward M, Nixon J, Gladding S. . An entrustable professional activity (EPA) for handoffs as a model for EPA assessment development. Acad Med. 2014; 89 10: 1335– 1340. [DOI] [PubMed] [Google Scholar]
  • 26. Chen HC, McNamara M, Teherani A, et al. Developing entrustable professional activities for entry into clerkship. Acad Med. 2016; 91 2: 247– 255. [DOI] [PubMed] [Google Scholar]
  • 27. Rekman J, Gofton W, Dudek N, et al. Entrustability scales: outlining their usefulness for competency-based clinical assessment. Acad Med. 2016; 91 2: 186– 190. [DOI] [PubMed] [Google Scholar]
  • 28. Crossley J, Johnson G, Booth J, et al. Good questions, good answers: construct alignment improves the performance of workplace-based assessment scales. Med Educ. 2011; 45 6: 560– 569. [DOI] [PubMed] [Google Scholar]
  • 29. Hauer KE, Boscardin C, Fulton TB, et al. Using a curricular vision to define entrustable professional activities for medical student assessment. J Gen Intern Med. 2015; 30 9: 1344– 1348. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30. Patton MQ. . Qualitative Evaluation and Research Methods. Beverly Hills, CA: SAGE Publications Inc; 1990. [Google Scholar]
  • 31. Jones J, Hunter D. . Consensus methods for medical and health services research. BMJ. 1995; 311 7001: 376– 380. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32. Van de Ven AH, Delbecq AL. . The nominal group as a research instrument for exploratory health studies. Am J Public Health. 1972; 62 3: 337– 342. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33. Pope C, Ziebland S, Mays N. . Qualitative research in health care: analysing qualitative data. BMJ. 2000; 320 7227: 114– 116. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34. Frank JR, Danoff D. . The CanMEDS initiative: implementing an outcomes-based framework of physician competencies. Med Teach. 2007; 29 7: 642– 647. [DOI] [PubMed] [Google Scholar]
  • 35. van der Vleuten CP. . The assessment of professional competence: developments, research and practical implications. Adv Health Sci Educ. 1996; 1 1: 41– 67. [DOI] [PubMed] [Google Scholar]
  • 36. Kane MT. . Validating the interpretations and uses of test scores. J Educ Meas. 2013; 50 1: 1– 73. http://onlinelibrary.wiley.com/doi/10.1111/jedm.12000/pdf. Accessed March 6, 2017. [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials


Articles from Journal of Graduate Medical Education are provided here courtesy of Accreditation Council for Graduate Medical Education

RESOURCES