Abstract
Introduction:
Competency-based education requires high-quality feedback to guide students’ acquisition of competencies. Sound assessment and feedback systems, such as ePortfolios, are needed to facilitate seeking and giving feedback during clinical placements. However, it is unclear whether the written feedback comments in ePortfolios are of high quality and aligned with the current competency focus. Therefore, this study investigates the quality of written feedback comments in ePortfolios of healthcare students, as well as how these feedback comments align with the CanMEDS roles.
Methods:
A qualitative textual analysis was conducted. 2,349 written feedback comments retrieved from the ePortfolios of 149 healthcare students (specialist medicine, general practice, occupational therapy, speech therapy and midwifery) were analysed retrospectively using deductive content analysis. Two structured categorisation matrices, one based on four literature-derived feedback quality criteria (performance, judgment, elaboration and improvement) and another one on the seven CanMEDS roles (Medical Expert, Communicator, Collaborator, Leader, Health Advocate, Scholar and Professional), guided the analysis.
Results:
The minority of the feedback comments (n = 352; 14.9%) could be considered of high quality because they met all four quality criteria. Most feedback comments were of moderate quality and met only two to three quality criteria. Regarding the CanMEDS roles, the Medical Expert role was most frequently represented in the feedback comments, as opposed to the roles Leader and Health Advocate.
Discussion:
The results highlighted that providing high-quality feedback is challenging. To respond to these challenges, it is recommended to set up individual and continuous feedback training.
Introduction
In healthcare education, societal pressures have driven a shift to competency-based education (CBE), in which graduate outcomes are formulated as competencies and are aligned with the roles graduates will play in the workplace [1]. A range of competency frameworks is available to organise competencies in a coherent manner [1]. The Canadian Medical Education Directions for Specialists (CanMEDS) competency framework is widely used and has been validated in the context of different healthcare professions [2]. The CanMEDS framework clusters competencies which are required by a competent professional into seven roles: Medical Expert, Communicator, Collaborator, Leader, Health Advocate, Scholar and Professional [3]. Students have great opportunity during clinical placements students to engage in these roles and to achieve the underlying competencies [4,5].
From a CBE perspective it is of crucial importance that during these clinical placements students receive high-quality feedback [6]. Therefore, sound assessment systems that facilitate the delivery of high-quality feedback are required [6]. This explains the adoption of ePortfolios, which support students in seeking feedback and teachers and clinical mentors in giving feedback [7]. A written feedback comment collected in an ePortfolio can be defined as “specific information about the comparison between a trainee’s observed performance and a standard, given with the intent to improve the trainee’s performance.” [8, p.193] As a result, written feedback comments can be considered valuable and valid data sources for students, teachers and clinical mentors. For students, written feedback comments are expected to guide their self-regulated learning process [9]. For teachers and clinical mentors, written feedback comments facilitate identifying struggling students and enable early remediation [10]. Furthermore, these written feedback comments support teachers and clinical mentors when making trustworthy and defensible decisions about the student’s progress [11].
For written feedback comments to be valuable for students, teachers and clinical mentors it is imperative that they are of high quality [12,13]. In accordance with van de Ridder et al.’s [8] definition of feedback, four distinct quality criteria can be used to characterize feedback comments as high-quality. Such feedback comments contain (1) specific information about the student’s performance [14], (2) a judgment about that performance [14], (3) elaboration on why the performance was judged that way [15] and (4) guidance on how the student’s performance can be improved [16]. These four feedback quality criteria align closely with the widely recognized feedback model introduced by Hattie and Timperley [17]. This model articulates three essential questions that effective feedback addresses, representing distinct facets of feedback: (1) ‘Where am I going?’—pertaining to the performance criterion (feed-up); (2) ‘How am I going?’—relating to judgment and elaboration criteria (feed-back); and (3) ‘Where to next?’—focused on the improvement criterion (feed-forward). In order to obtain sufficiently rich written feedback comments to make valid decisions about student performance [18], it is imperative that feedback comments meet all four established quality criteria. Of course, the individual criteria can be present in subsections of feedback comments. However, when these subsections are combined, the four feedback quality criteria are expected to be represented.
In the context of CBE, an additional focus on feedback quality is needed that requires written feedback comments to develop a picture of the student’s competency acquisition [19]. Thus, written feedback comments must align with the roles and underlying competencies that graduating healthcare professionals should develop in view of providing effective and safe patient care, as specified in competency frameworks (e.g., CanMEDS framework) [20].
ePortfolios surged in popularity to support feedback provision during clinical placements with an emphasis on high-quality written feedback comments to establish CBE. However, empirical research about the quality of written feedback comments in ePortfolios is lacking. Few studies have investigated the quality of written feedback comments in students’ ePortfolios and have, to the best of our knowledge, only focused on feedback quality criteria [21,22,23]. The present study adopts a double focus by not only looking at the quality of written feedback comments in ePortfolios (1), but by also investigating how these written feedback comments are aligned with the seven CanMEDS roles (2). To achieve this, two research questions were addressed:
RQ1. What is the quality of written feedback comments in ePortfolios of healthcare students?
RQ2. How are the written feedback comments in these ePortfolios aligned with the CanMEDS roles?
Methods
Design
We conducted a qualitative, textual analysis study [24] by carrying out a retrospective analysis of written feedback comments collected in the ePortfolios of healthcare students during clinical placements. In this respect, we selected the constructivist research paradigm. Within this paradigm, it is argued that it is impossible for researchers to separate themselves from their beliefs and values [25]. Therefore, in the section ‘Reflexivity’ below, we describe our backgrounds to clarify how this may have influenced our interpretation and coding of the feedback comments.
Ethical approval
Ethical approval was obtained from the Ethical Committee of the Faculty of Psychology and Educational Sciences of Ghent University (reference #2021-34) and a Data Transfer Agreement was signed between the Medbook company and Ghent University.
Context
The study was conducted in Flanders (Belgium) and the feedback comments analysed were written in Dutch. Since feedback quality criteria and the CanMEDS roles are not dependent on the nature of a specific healthcare profession [2], we included different educational programs in this study. Five healthcare educational programs participated: specialist medicine (post-graduate), general practice (post-graduate), occupational therapy (undergraduate), speech therapy (undergraduate) and midwifery (undergraduate). As such, we were able to develop a general picture of the feedback comments in ePortfolios across educational programs, both in terms of quality and alignment with the CanMEDS roles.
The included educational programs implemented the same ePortfolio platform (Medbook) to support feedback provision during clinical placements. Because of the availability of a large number of feedback comments in Medbook’s database, we collaborated with the Medbook company to include the feedback comments as research data in this study.
Data collection
The feedback comments used as research data in this study were written in the students’ ePortfolios in an open text box as feedback on a reflection (undergraduate programs) or as part of a (low-stakes) assessment (postgraduate programs). The frequency with which feedback comments were written varied by educational program. The data collection was conducted in June 2021.
To ensure participants’ privacy, we took three preventive measures. First, students’ explicit consent was required to participate in the study, which included allowing the Medbook company to share their feedback comments with the researchers. All students from the participating educational programs taking an internship during the academic year 2020–2021 were invited to participate. Second, this information was provided to students via a pop-up window in Medbook, ensuring that we did not need personal information to contact them. When a student logged into Medbook, this pop-up window appeared. The Medbook company tracked which students agreed to participate and then extracted their data from the Medbook database. Third, before providing the data to the researchers, the Medbook company replaced all names included in the feedback comments with the placeholder ‘[anonymised]’. We received one dataset for each educational program containing the feedback comments and the dates at which they were written.
Data analysis
The feedback comments were analysed using deductive content analysis, a commonly used method for analysing written texts to describe and quantify phenomena (e.g. feedback comments) [26,27]. We followed the three phases of the content analysis process described by Elo and Kyngäs [27]: (1) preparation, (2) organising and (3) reporting.
Preparation phase
First, we selected the unit of analysis, which was a complete feedback comment extracted from the ePortfolios. Subsequently, we read through these feedback comments to familiarise ourselves with the data.
Organising phase
During the organising phase, we developed two structured categorisation matrices: one based on the four literature-derived quality criteria (performance, judgment, elaboration and improvement) and one based on the seven CanMEDS roles (Medical Expert, Communicator, Collaborator, Leader, Health Advocate, Scholar and Professional) with underlying competencies (see Appendix A and B). We obtained permission to use the CanMEDS framework from the Royal College of Physicians and Surgeons of Canada. To test the transparency of these categorisation matrices, they were used by three researchers (SVO, ME, HD), independently of one another, to code a pilot set of 30 feedback comments. Afterwards, the researchers discussed coding discrepancies and iteratively refined the categorisation matrices through four additional review rounds. After deciding on the final categorisation matrices, the feedback comments were coded in two stages.
In the first stage (July 2021), the feedback comments were coded in the annotation platform INCEpTION [28] using the four quality criteria. To ensure coding reliability, two researchers (SVO and SJA) coded 100 feedback comments independent of one another (see Table 1 for the Cohen’s Kappa values). Afterwards, these double-coded feedback comments were discussed to identify and resolve discrepancies. Here, special attention was paid to the criterion elaboration as the Cohen’s Kappa value for this code indicated fair agreement among the researchers [29]. We made minor changes to clarify differences between coding categories and identified exemplary instances that guided the subsequent coding. The first author then independently coded the complete set of feedback comments.
Table 1.
Cohen’s Kappa values.
|
| |
|---|---|
| CODE | COHEN’S KAPPA VALUE |
|
| |
| Quality criteria | |
|
| |
| Performance | 0.77 |
|
| |
| Judgment | 0.68 |
|
| |
| Elaboration | 0.24 |
|
| |
| Improvement | 0.83 |
|
| |
| CanMEDS roles | |
|
| |
| Medical Expert | 0.63 |
|
| |
| Communicator | 0.62 |
|
| |
| Collaborator | 0.60 |
|
| |
| Leader | 0.51 |
|
| |
| Health Advocate | 0.50 |
|
| |
| Scholar | 0.63 |
|
| |
| Professional | 0.21 |
|
| |
In the second stage (June 2022), the text coded with the quality criteria performance, elaboration and improvement was further enriched with the seven CanMEDS roles in Microsoft Excel (Office 365). The text coded with the judgment criterion was not considered because it referred to adjectives that per se cannot be related to the CanMEDS roles. We also ensured reliability in this stage by double-coding 100 feedback comments (SVO and OJ) (see Table 1 for the Cohen’s Kappa values). Discrepancies in the double-coded feedback comments were resolved through discussion. The roles Professional, Collaborator, Leader and Health Advocate were given careful attention because agreement among coders on these roles was fair to moderate [29]. Next, the first author coded independently all the feedback comments and subsequently discussed certain interpretations and uncertainties with the research team to reach a consensus.
Reflexivity
The first author, SVO, is a PhD student in Educational Sciences and is entirely independent from the educational programs in which the feedback comments were collected. The students had no personal connection with her. SVO tested the transparency of the categorisation matrices in collaboration with ME and HD. ME is a postdoctoral researcher with a background in midwifery and 20 years of experience as the coordinator of the midwifery program. HD is a researcher and also a midwife, providing her with firsthand experience in writing feedback comments to students. Their backgrounds offer unique perspectives on the quality of the feedback comments. SJA is a Master’s student in Educational Studies, and OJ is a PhD student in Educational Studies and Health Sciences. They received comprehensive training to conduct double coding. Throughout the process, discussions were held at various time points to address any uncertainties and discrepancies.
Results
Of the 3,504 students who received the pop-up window in Medbook, 149 students gave consent to share the feedback comments in their ePortfolios. This resulted in a set of 2,349 feedback comments. Table 2 provides an overview of the number of students that gave consent and the number of analysed feedback comments for each educational program.
Table 2.
Overview of the number of participating students and analysed feedback comments per educational program.
|
| ||
|---|---|---|
| EDUCATIONAL PROGRAM | NO. OFSTUDENTS (%) | NO. OFFEEDBACK COMMENTS (%) |
|
| ||
| Specialistic medicine | 47 (31.54) | 649 (27.63) |
|
| ||
| General practice | 62 (41.61) | 256 (10.90) |
|
| ||
| Occupational therapy | 9 (6.04) | 229 (9.75) |
|
| ||
| Speech therapy | 8 (5.37) | 204 (8.68) |
|
| ||
| Midwifery | 23 (15.44) | 1,011 (43.04) |
|
| ||
| Total | 149 (100) | 2,349 (100) |
|
| ||
Quality of written feedback comments
During the first coding stage, we coded the feedback comments by considering the four quality criteria (See Appendix C for an example). Overall, most feedback comments fulfilled the criteria performance (n = 1,681; 71.56%), judgment (n = 1,613; 68.67%) and improvement (n = 1,298; 55.26%). The criterion elaboration was mostly lacking. This criterion was observed in less than a quarter of the feedback comments (n = 543; 23.12%).
Based on the number of coded quality criteria in a feedback comment these can be organised at three quality levels: low, moderate and high quality. We now describe the characteristics of the feedback comments from the different quality levels.
Low-quality feedback comments
Low-quality feedback comments met none (n = 342; 14.56%) or one (n = 347; 14.77%) quality criterion. Almost one-third of the feedback comments were of low quality (n = 689; 29.33%). Only a small number of the low-quality feedback comments contained information related to the performance (n = 27; 3.92%), the judgment (n = 8; 1.16%) or the elaboration criterion (n = 2; 0.29%). The criterion that was most often represented was improvement (n = 310; 44.99%), as illustrated in the following examples:
‘Tighten skin when you insert the needle.’ [comment_145, specialist medicine]
‘Keep focus from start to finish. A small mistake can have big consequences.’ [comment_516, specialist medicine]
Moderate-quality feedback comments
Feedback comments of moderate quality met two (n = 544; 23.16%) or three (n = 764; 32.52%) quality criteria. The analysis revealed that the quality of most of the feedback comments was moderate (n = 1,308; 55.68%). In contrast to low-quality feedback comments, almost all moderate-quality feedback comments applied the criteria performance (n = 1,302; 99.54%) and judgment (n = 1,253; 95.80%). Almost half of the feedback comments included information related to the criterion improvement (n = 636; 48.62%). The elaboration criterion was often lacking, as it was represented in only a few feedback comments (n = 189; 14.45%). The following feedback comments met the criteria performance, judgment and improvement, but missed information related to the elaboration criterion:
‘Pleasant (non)verbal communication. Multiple problems in 1 consultation, handled this fairly flexibly. Make sure things are more prepared before they [the patients] are in the chair.’ [comment_182, general practice]
‘Wound care: Technique is known. Take more fluid and dare to clean the wound seam more thoroughly. Always hold the skin back when removing the sticker!!! Also again observe the patient more while acting.’ [comment_1358, midwifery]
Hig-Hquality feedback comments
When feedback comments met all four quality criteria, they were of high quality. Only a minority of the feedback comments were rated as high quality (n = 352; 14.99%). The following examples of feedback comments met all four quality criteria:
‘[anonymised] has been posting diabolos smoothly, also duravent and T-tube. Goes smoothly, doesn’t take long. Still enough attention to place the diabolo properly antero-inferiorly, though.’ [comment_454, specialist medicine]
‘Positive feedback from the supervising occupational therapist. You had prepared the activity well which also enabled you to differentiate. Tip: You can also organise this activity with fly swatters and a balloon.’ [comment_712, occupational therapy]
Alignment of written feedback comments with the CanMEDS roles
In the second coding stage, we further enriched the feedback comments with the CanMEDS roles. An example feedback comment for each CanMEDS role is provided in Appendix C.
The analysis indicated that the role Medical Expert was identified most frequently in the feedback comments (n = 1,530; 65.13%). The roles Communicator (n = 790; 33.63%), Scholar (n = 712; 30.31%), Collaborator (n = 634; 26.99%) and Professional (n = 375; 15.96%) were less frequently represented. The roles Leader (n = 256; 10.90%) and Health Advocate (n = 217; 9.24%) appeared in the least number of feedback comments.
Furthermore, it is important to note that a minor section of the feedback comments could not be related to any CanMEDS role (n = 354; 15.07%). Examples of such instances are:
‘No remarks’ [comment_169, specialist medicine]
‘Practice makes perfect!’ [comment_2, specialist medicine]
‘Keep up the good work!’ [comment_893, general practice]
Discussion
To our knowledge, this is the first study to investigate the quality of written feedback comments in ePortfolios, as well as how these feedback comments can be aligned with the CanMEDS roles. The results revealed that the minority of the feedback comments could be considered of high quality. Most feedback comments met two or three quality criteria and were of moderate quality. Furthermore, the Medical Expert role was most frequently represented in the feedback comments, in contrast to the roles Leader and Health Advocate.
The results confirm observations about the critical quality of written feedback comments in healthcare education in general [30] and in ePortfolios in particular [21,23]. Most of the feedback comments contained information related to the criteria performance, judgment and improvement. However, the elaboration criterion was often lacking, which implies that crucial information making the feedback comments specific, was mostly missing. The paucity of specific feedback comments is a well-known phenomenon in healthcare education [14,23,31,32] and raises concerns about feedback providers’ readiness and skills to provide feedback. While acknowledging their responsibility to provide students with high-quality feedback to foster competency development and performance improvement, feedback providers in healthcare education face considerable challenges in meeting this demand [33,34]. The provision of high-quality feedback necessitates a diverse skill set among feedback providers [17]. Research reveals a deficiency in feedback skills among feedback providers in healthcare education [35], as they struggle to use feedback forms accurately, encounter difficulties in applying predefined learning outcomes as assessment criteria for students’ competencies and face challenges in providing high-quality feedback, even after receiving training [35,36,37]. As the provision of non-specific feedback comments has a negative impact on the student’s feedback-seeking behaviour [13], it is imperative for future research to investigate why feedback providers struggle to provide specific feedback to students and to identify the precise skills that require development to achieve feedback of a higher quality.
This study sheds new light on the representation of the CanMEDS roles in written feedback comments. The results show that most feedback was given in relation to the Medical Expert role. This can be explained because healthcare professionals consider this as the most important role; thus leading to a biased feedback orientation when assessing students [38]. The roles Leader and Health Advocate were represented to a lesser extent. This might be because these roles are underrepresented in healthcare curricula and learning activities during clinical placements [39,40]. Furthermore, teachers and clinical mentors know little about how to teach and assess these roles [41]. Although opportunities to learn about the Leader and Health Advocate role are optimal during clinical placements, they do not seem to be adequately reflected in feedback comments, which hinders related acquisition of these roles [42]. Given the inconsistent representation of certain CanMEDS roles in feedback comments [43], setting up training initiatives to improve awareness of the CanMEDS roles and to provide practical ideas and a shared language to incorporate the roles in feedback comments, is recommended [12,41,42].
Enhancing the quality of written feedback comments seems to be a Sisyphean task [44]. Training initiatives are available to train teachers and clinical mentors in writing high-quality feedback comments. Research about generic training initiatives – such as formal training programs and workshops – mirror inconsistent effects [31,45], mainly due to competing professional interests for time, including patient care, teaching, research, faculty promotion and administration [32]. As an alternative, a personalised approach could be adopted that focuses on giving feedback to feedback providers on their feedback comments and on coaching over a longer period of time [31]. Certainly in the context of ePortfolio use, an individual, continuous training approach is recommended [46]. However, the implementation of such a personal training approach in a traditional way (e.g., feedback on feedback from educational staff) is challenging due to its time and resource-intensive nature. Future research is required to explore how these time and resource constraints can be overcome.
The present study demonstrated that the manual analysis of written feedback comments is time-consuming. This is a growing problem as more and more digital feedback tools are being adopted in healthcare education, leading to increasing amounts of written feedback comments [47]. Recent technological advances in the field of artificial intelligence and Natural Language Processing (NLP) might be helpful to evaluate large amounts of feedback comments in a short amount of time [48]. Researchers already demonstrated the potential utility of NLP techniques to classify the quality of written feedback comments [47,48]. Future research should explore how these NLP techniques could be integrated into feedback tools in a way that they can provide real-time support to teachers and clinical mentors in writing feedback comments.
This study should be considered in light of its limitations. First, we only analysed written feedback comments. It is possible that more high-quality verbal feedback was provided to the students during formal or informal feedback conversations, but that this was not documented in the ePortfolios [49]. Given that this study aimed to investigate written feedback comments in ePortfolios, this does not impact the results. However, future research could explore whether the results of the current analysis are also applicable to verbal feedback. Secondly, we did not consider individual differences in student’s, teachers’ and clinical mentors’ characteristics. Mooney et al. [50], for example, found that feedback quality was associated with the feedback provider’s gender. Additionally, a within-subject variation is possible, depending on the feedback performance phase being focused upon by the same feedback provider. However, to ensure the participants’ privacy, this was not questioned when collecting the feedback comments. A third limitation is linked to the differences in sample sizes for each educational program. This can be explained by the timing of data collection. Due to an extended ethical approval process, data were collected later than initially planned (June 2021). This period coincided with the end of the academic year, resulting in fewer students who were still on their clinical placements and consequently fewer students opening their ePortfolios and viewing the pop-up window. Although this limitation is not a shortcoming for the present study, this could be an issue when subsequent research would focus on the quality of feedback comments in each educational program. The current dataset cannot be used to compare the feedback comments of different educational programs.
In conclusion, the results of this study highlighted that providing high-quality feedback that develops a picture of the student’s competency acquisition is challenging. The majority of the feedback comments were of moderate quality and focused on the student’s performance related to the Medical Expert role. To respond to these challenges, it is recommended to set up individual, continuous training initiatives. In this regard, the opportunities to use NLP techniques to offer real-time support to feedback providers in writing feedback should be explored.
Data Accessibility Statement
The participants of this study did not give written consent for their data to be shared publicly, so due to the sensitive nature of the research supporting data is unavailable.
Additional Files
The additional files for this article can be found as follows:
Structured categorisation matrix quality criteria.
Structured categorisation matrix CanMEDS roles.
Example feedback comment for each feedback quality criterion and CanMEDS role (translated from Dutch).
Acknowledgements
The authors would like to thank Karel Verbert (Medbook) for his cooperation and help during data collection. The authors also thank Oona Janssens, Helena Demey and Sarah-Jane Arschoot for their help during the organising phase of the data analysis.
Funding Statement
This work was supported by Research Foundation Flanders (FWO, Strategic Basic Research (SBO) under Grant S003219N.
Funding Information
This work was supported by Research Foundation Flanders (FWO, Strategic Basic Research (SBO) under Grant S003219N.
Competing Interests
The authors have no competing interests to declare.
References
- 1.Holmboe ES, Sherbino J, Englander R, Snell L, Jason R. A call to action: the controversy of and rationale for competency-based medical education. Med Teach. 2017; 39(6): 574–81. DOI: 10.1080/0142159X.2017.1315067 [DOI] [PubMed] [Google Scholar]
- 2.Janssens O, Embo M, Valcke M, Haerens L. An online Delphi study to investigate the completeness of the CanMEDS roles and the relevance, formulation, and measurability of their key competencies within eight healthcare disciplines in Flanders. BMC Med Educ. 2022; 22(260): 1–14. DOI: 10.1186/s12909-022-03308-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Frank JR. CanMEDS: a framework for teaching and assessing competencies. In: Sherbino J, Frank JR (Eds.) Educational design a CanMEDS guide for the health professions. Ottawa: Royal College of Physicians and Surgeons; 2011. pp. 17–22. [Google Scholar]
- 4.Rodger S, Webb G, Devitt L, et al. A clinical education and practice placements in the allied health professions: an international perspective. J Allied Health. 2008; 37(1): 53–62. [PubMed] [Google Scholar]
- 5.Steven K, Wenger E, Boshuizen H, Scherpbier A, Dornan T. How clerkship students learn from real patients in practice settings. Acad Med. 2014; 89(3): 469–76. DOI: 10.1097/ACM.0000000000000129 [DOI] [PubMed] [Google Scholar]
- 6.Holmboe ES, Sherbino J, Long DM, Swing SR, Frank JR. The role of assessment in competency-based medical education. Med Teach. 2010; 32(8): 676–82. DOI: 10.3109/0142159X.2010.500704 [DOI] [PubMed] [Google Scholar]
- 7.Janssens O, Haerens L, Valcke M, et al. The role of ePortfolios in supporting learning in eight healthcare disciplines: a scoping review. Nurse Educ Pract. 2022; 63: 1–12. DOI: 10.1016/j.nepr.2022.103418 [DOI] [PubMed] [Google Scholar]
- 8.Van De Ridder JMM, Stokking KM, McGaghie WC, Ten Cate OTJ. What is feedback in clinical education? Med Educ. 2008; 42(2): 189–97. DOI: 10.1111/j.1365-2923.2007.02973.x [DOI] [PubMed] [Google Scholar]
- 9.Ginsburg S, Watling CJ, Schumacher DJ, Gingerich A, Hatala R. Numbers encapsulate, words elaborate: toward the best use of comments for assessment and feedback on entrustment ratings. Acad Med. 2021; 96(7): 81–86. DOI: 10.1097/ACM.0000000000004089 [DOI] [PubMed] [Google Scholar]
- 10.Tomiak A, Braund H, Egan R, et al. Exploring how the new entrustable professional activity assessment tools affect the quality of feedback given to medical oncology residents. J Cancer Educ. 2020; 35(1): 165–77. DOI: 10.1007/s13187-018-1456-z [DOI] [PubMed] [Google Scholar]
- 11.Lefebvre C, Hiestand B, Glass C, et al. Examining the effects of narrative commentary on evaluators’ summative assessments of resident performance. Eval Heal Prof. 2020; 43(3): 159–61. DOI: 10.1177/0163278718820415 [DOI] [PubMed] [Google Scholar]
- 12.Renting N, Gans ROB, Borleffs JCC, et al. A feedback system in residency to evaluate CanMEDS roles and provide high-quality feedback: exploring its application. Med Teach. 2016; 38(7): 738–45. DOI: 10.3109/0142159X.2015.1075649 [DOI] [PubMed] [Google Scholar]
- 13.Fu RH, Cho YH, Quattri F, Monrouxe LV. ‘I did not check if the teacher gave feedback’: a qualitative analysis of Taiwanese postgraduate year 1 trainees’ talk around e-portfolio feedback-seeking behaviours. BMJ Open. 2019; 9(1): 1–9. DOI: 10.1136/bmjopen-2018-024425 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Canavan C, Holtman MC, Richmond M, Katsufrakis PJ. The quality of written comments on professional behaviors in a developmental multisource feedback program. Acad Med. 2010; 85(10): 106–9. DOI: 10.1097/ACM.0b013e3181ed4cdb [DOI] [PubMed] [Google Scholar]
- 15.Ginsburg S, Regehr G, Lingard L, Eva KW. Reading between the lines: faculty interpretations of narrative evaluation comments. Med Educ. 2015; 49(3): 296–306. DOI: 10.1111/medu.12637 [DOI] [PubMed] [Google Scholar]
- 16.Ginsburg S, Van Der Vleuten CPM, Eva KW. The hidden value of narrative comments for assessment: a quantitative reliability analysis of qualitative data. Acad Med. 2017; 92(11): 1617–21. DOI: 10.1097/ACM.0000000000001669 [DOI] [PubMed] [Google Scholar]
- 17.Hattie J, Timperley H. The power of feedback. Rev. Educ. Res. 2007; 77(1): 81–112. DOI: 10.3102/003465430298487 [DOI] [Google Scholar]
- 18.Cook DA, Brydges R, Ginsburg S, Hatala R. A contemporary approach to validity arguments: A practical guide to Kane’s framework. Med Educ. 2015; 49(6): 560–75. DOI: 10.1111/medu.12678 [DOI] [PubMed] [Google Scholar]
- 19.de Jong LH, Bok HGJ, Schellekens LH, et al. Shaping the right conditions in programmatic assessment: how quality of narrative information affects the quality of high-stakes decision-making. BMC Med Educ. 2022; 22(409): 1–10. DOI: 10.1186/s12909-022-03257-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Ginsburg S. Hidden in plain sight: the untapped potential of written assessment comments. PhD thesis. Maastricht University, Maastricht. 2016. DOI: 10.26481/dis.20160901sg [DOI] [Google Scholar]
- 21.Barrett A, Galvin R, Steinert Y, et al. Profiling postgraduate workplace-based assessment implementation in Ireland: a retrospective cohort study. Springerplus. 2016; 5(133): 1–7. DOI: 10.1186/s40064-016-1748-x [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Bleasel J, Burgess A, Weeks R, Haq I. Feedback using an ePortfolio for medicine long cases: quality not quantity. BMC Med Educ. 2016; 16(278): 1–11. DOI: 10.1186/s12909-016-0801-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Tham TCK, Burr B, Boohan M. Evaluation of feedback given to trainees in medical specialties. Clin Med J R Coll Physicians London. 2017; 17(4): 303–6. DOI: 10.7861/clinmedicine.17-4-303 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Park S, Griffin A, Gill D. Working with words: exploring textual analysis in medical education research. Med Educ. 2012; 46(4): 372–80. DOI: 10.1111/j.1365-2923.2011.04184.x [DOI] [PubMed] [Google Scholar]
- 25.Kivunja C, Kuyini, AB. Understanding and Applying Research Paradigms in Educational Contexts. Int. J. High. Educ. 2017; 6(5): 26–41. DOI: 10.5430/ijhe.v6n5p26 [DOI] [Google Scholar]
- 26.Cole FL. Content analysis: process and application. Clin nurse Spec. 1988; 2(1): 53–57. DOI: 10.1097/00002800-198800210-00025 [DOI] [PubMed] [Google Scholar]
- 27.Elo S, Kyngäs H. The qualitative content analysis process. J Adv Nurs. 2008; 62(1): 107–15. DOI: 10.1111/j.1365-2648.2007.04569.x [DOI] [PubMed] [Google Scholar]
- 28.Klie J-C, Bugert M, Boullosa B, de Castilho RE, Gurevych I. The INCEpTION Platform: machine-assisted and knowledge-oriented interactive annotation. In: Zhao D (ed.). Proceedings of the 27th international conference on computational linguistics: system demonstrations, 20–26 Aug 2018, Santa Fe, New Mexico, USA. Stroudsburg (PA): Association for Computational Linguistics; 2018. 5–9. [Google Scholar]
- 29.Cohen J. A coefficient of agreement for nominal scales. Educ Psychol Meas. 1960; 20: 37–46. DOI: 10.1177/001316446002000104 [DOI] [Google Scholar]
- 30.Raaum SE, Lappe K, Colbert-Getz JM, Milne CK. Milestone implementation’s impact on narrative comments and perception of feedback for internal medicine residents: a mixed methods study. J Gen Intern Med. 2019; 34(6): 929–35. DOI: 10.1007/s11606-019-04946-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Govaerts MJB, van de Wiel MWJ, van der Vleuten CPM. Quality of feedback following performance assessments: does assessor expertise matter? Eur J Train Dev. 2013; 37(1): 105–25. DOI: 10.1108/03090591311293310 [DOI] [Google Scholar]
- 32.Roshan A, Wagner N, Acai A, et al. Comparing the quality of narrative comments by rotation setting. J Surg Educ. 2021; 78(6): 2070–77. DOI: 10.1016/j.jsurg.2021.06.012 [DOI] [PubMed] [Google Scholar]
- 33.Davila-Cervantes A, Foulds, JL, Gomaa, NA, Rashid, M. Experiences of Faculty Members Giving Corrective Feedback to Medical Trainees in a Clinical Setting. J Contin Educ Health Prof. 2021; 41(1): 24–30. DOI: 10.1097/CEH.0000000000000322 [DOI] [PubMed] [Google Scholar]
- 34.Kim B, Rajagopalan A, Tabasky EM, Reddy SS, Topor DR. Sharing perspectives on feedback: a combined resident-faculty workshop. BMC Med Educ. 2022; 22: 1–8. DOI: 10.1186/s12909-022-03253-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Nugraheny E, Claramita M, Rahayu G, Kumara A. Feedback in the nonshifting context of the midwifery clinical education in Indonesia: A mixed methods study. Iran J Nurs Midwifery Res. 2016; 21(6): 628–34. DOI: 10.4103/1735-9066.197671 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Bing-You R, Varaklis K, Hayes V, et al. The feedback tango: An integrative review and analysis of the content of the teacher-learner feedback exchange. Acad Med. 2018; 93(4): 657–63. DOI: 10.1097/ACM.0000000000001927 [DOI] [PubMed] [Google Scholar]
- 37.Christiansen B, Averlid G, Baluyot C, et al. Challenges in the assessment of nursing students in clinical placements: Exploring perceptions among nurse mentors. Nurs Open. 2021; 8(3): 1069–76. DOI: 10.1002/nop2.717 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Li P, Jiang F, Yin L, et al. Perceptions of the CanMEDS competencies of faculty and students in different curriculum systems of a medical school in China. Adv Med Educ Pract. 2022; 13: 1061–70. DOI: 10.2147/AMEP.S367129 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Bugaj TJ, Schmid C, Koechel A, et al. Shedding light into the black box: A prospective longitudinal study identifying the CanMEDS roles of final year medical students’ on-ward activities. Med Teach. 2017; 39(8): 883–90. DOI: 10.1080/0142159X.2017.1309377 [DOI] [PubMed] [Google Scholar]
- 40.Binnendyk J, Pack R, Field E, Watling C. Not wanted on the voyage: highlighting intrinsic CanMEDS gaps in Competence by Design curricula. Can. Med. Educ. J. 2021; 12(4): 39–47. DOI: 10.36834/cmej.70950 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Puddester D, MacDonald CJ, Clements D, Gaffney J, Wiesenfeld L. Designing faculty development to support the evaluation of resident competency in the intrinsic CanMEDS roles: practical outcomes of an assessment of program director needs. BMC Med Educ. 2015; 15(100): 1–9. DOI: 10.1186/s12909-015-0375-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Renting N, Raat ANJ, Dornan T, et al. Integrated and implicit: how residents learn CanMEDS roles by participating in practice. Med Educ. 2017; 51(9): 942–52. DOI: 10.1111/medu.13335 [DOI] [PubMed] [Google Scholar]
- 43.Renting N, Dornan T, Gans ROB, et al. What supervisors say in their feedback: construction of CanMEDS roles in workplace settings. Adv in Health Sci Educ. 2016; 21: 375–87. DOI: 10.1007/s10459-015-9634-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Dudek NL, Marks MB, Wood TJ, et al. Quality evaluation reports: can a faculty development program make a difference? Med Teach. 2012; 34(11): e725–31. DOI: 10.3109/0142159X.2012.689444 [DOI] [PubMed] [Google Scholar]
- 45.Pelgrim EA, Kramer AW, Mokkink HG, Van Der Vleuten CP. Quality of written narrative feedback and reflection in a modified mini-clinical evaluation exercise: an observational study. BMC Med Educ. 2012; 12(97): 1–6. DOI: 10.1186/1472-6920-12-97 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Van Ostaeyen S, Embo M, Schellens T, Valcke M. Training to support ePortfolio users during clinical placements: a scoping review. Med Sci Educ. 2022; 32(4): 921–28. DOI: 10.1007/s40670-022-01583-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47.Ötleş E, Kendrick DE, Solano QP, et al. Using natural language processing to automatically assess feedback quality: findings from 3 surgical residencies. Acad Med. 2021; 96(10): 1457–60. DOI: 10.1097/ACM.0000000000004153 [DOI] [PubMed] [Google Scholar]
- 48.Marcotte KM, Ötleş E, Thelen AE, et al. Using natural language processing to determine factors associated with high-quality feedback. Glob Surg Educ – J Assoc Surg Educ. 2022; 1(58): 1–5. DOI: 10.1007/s44186-022-00051-y [DOI] [Google Scholar]
- 49.Elamin A, Obeidat M, Davis G. The ePortfolio in UK cardiology training: time for a new digital platform? Br J Cardiol. 2021; 28(3): 1–4. DOI: 10.5837/bjc.2021.031 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.Mooney CJ, Pascoe JM, Blatt AE, et al. Predictors of faculty narrative evaluation quality in medical school clerkships. Med Educ. 2022; 56(12): 1223–31. DOI: 10.1111/medu.14911 [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Structured categorisation matrix quality criteria.
Structured categorisation matrix CanMEDS roles.
Example feedback comment for each feedback quality criterion and CanMEDS role (translated from Dutch).
Data Availability Statement
The participants of this study did not give written consent for their data to be shared publicly, so due to the sensitive nature of the research supporting data is unavailable.
