Abstract
Background
The most recent iteration of the Accreditation Council for Graduate Medical Education (ACGME) duty hour regulations includes language mandating handoff education for trainees and assessments of handoff quality by residency training programs. However, there are a lack of validated tools for the assessment of handoff quality and to utilize for trainee education.
Methods
Faculty at two sites (University of Chicago and Yale) were recruited to participate in a workshop on handoff education. Video-based scenarios were developed to represent varying levels of performance in the domains of communication, professionalism and setting. Videos were shown in a random order and faculty were instructed to use the Handoff mini-CEX, a paper-based instrument with qualitative anchors defining each level of performance, to rate the handoffs.
Results
Forty-seven faculty members (14=site 1; 33=site 2) participated in the validation workshops providing a total of 172 observations (of a possible 191 (96%)). Reliability testing revealed a Cronbach’s alpha of 0.81 and Kendall’s coefficient of concordance of 0.59 (>0.6=high reliability). Faculty were able to reliably distinguish the different levels of performance in each domain in a statistically significant fashion (i.e. unsatisfactory professionalism mean 2.42 vs. satisfactory professionalism 4.81 vs. superior professionalism 6.01, p<0.001 trend test). Two-way ANOVA revealed no evidence of rater bias.
Conclusions
Using standardized video-based scenarios highlighting differing levels of performance, we were able to demonstrate evidence that the Handoff mini-CEX can draw reliable and valid conclusions regarding handoff performance. Future work to validate the tool in clinical settings is warranted.
Introduction
Over the last decade, there has been an unprecedented focus on physician handoffs in American hospitals. One major reason for this are the reductions in residency duty hours that have been mandated by the American Council for Graduate Medical Education (ACGME), first in 2003, and subsequently revised in 2011.(1)(2) As residents work fewer hours, experts believe that potential safety gains from reduced fatigue are countered by an increase in the number of handoffs, which represent a risk due to the potential miscommunication. Prior studies show that critical patient information is often lost or altered during this transfer of clinical information and professional responsibility, which can result in patient harm.(3)(4) As a result of these concerns, the ACGME now requires residency programs to “ensure and monitor effective, structured hand-over processes to facilitate both continuity of care and patient safety. Programs must ensure that residents are competent in communicating with team members in the hand-over process.”(2) Moreover, handoffs have also been a major improvement focus for organizations with broader scope than teaching hospitals, including the World Health Organization, Joint Commission, and the Society for Hospital Medicine (SHM).(5, 6, 7)
Despite this focus on handoffs, monitoring quality of handoffs has proven challenging due to lack of a reliable and validated tool to measure handoff quality. More recently, the Accreditation Council of Graduate Medical Education’s introduction of the Next Accreditation System with its focus on direct observation of clinical skills to achieve milestones makes it crucial for residency educators to have valid tools to measure competence in handoffs. As a result, it is critical that instruments to measure hand-off performance are not only created but also validated.(8)
To help fill this gap, we have previously reported on the development of a 9-item Hand-Off Clinical Examination Exercise (CEX) assessment tool. The Handoff CEX, designed for use by those participating in the handoff or by a third-party observer, can be used to rate the quality of patient hand-offs in domains such as professionalism and communication skills between the receiver and sender of patient information. (9, 10) Despite prior demonstration of feasibility of use, the initial tool was perceived as lengthy and redundant. In addition, while the tool has been shown to discriminate between performance of novice and expert nurses, the construct validity of this tool has not been established.(11) Establishing construct validity is important to ensuring that the tool can measure the construct in question, namely whether it detects those who are actually competent to perform handoffs safely and effectively. We present here the results of the development of a shorter mini-Handoff CEX along with the formal establishment of its’ construct validity, namely its ability to distinguish between levels of performance in three domains of handoff quality.
Methods
Adaption of Handoff CEX and development of the abbreviated tool
The 9-item Hand-off CEX is a paper-based instrument that was created by the investigators (LIH, JMF, VMA) to evaluate either the sender or the receiver of handoff communication and has been used in prior studies (see Appendix 1).(9, 10). The evaluation may be conducted by either an observer or by one of the hand-off participants. The instrument includes six domains: (1) setting (2) organization and efficiency (3) communication skills (4) content (5) clinical judgment and (6) humanistic skills/professionalism. Each domain is graded on a nine point rating scale, modeled on the widely used Mini-CEX (Clinical Evaluation Exercise) for real-time observation of clinical history and exam skills in internal medicine clerkships and residencies [1–3= unsatisfactory; 4–6= marginal/satisfactory; 7–9= superior].(12) This familiar 9-point scale is utilized in graduate medical education evaluation of the ACGME core competencies.
To standardize the evaluation, the instrument uses performance-based anchors for evaluating both the sender and the receiver of the hand-off information. The anchors are derived from functional evaluation of the roles of senders and receivers in our preliminary work at both the University of Chicago and Yale, best practices in other high-reliability industries, guidelines from the Joint Commission and the Society of Hospital Medicine and prior studies of effective communication in clinical systems.(5, 6, 13)
After piloting the Handoff CEX with the University of Chicago’s Internal Medicine residency program (n=280 handoff evaluations), a strong correlation was noted between the measures of content (medical knowledge), patient care, clinical judgment, organization/efficiency and communication skills. Moreover, the Handoff CEX’s Cronbach’s alpha, or measurement of internal reliability and consistency, was very high (alpha=0.95). Given the potential of redundant items, and in order to increase ease of use of the instrument, factor analysis was used to reduce the instrument to yield a shorter 3-item tool, the Handoff Mini-CEX, that assessed 3 of the initial items: setting, communication skills and professionalism. Overall, performance on these 3 items were responsible for 82% of the variance of overall sign-out quality (see Appendix 2).
Establishing Construct Validity of the mini-Handoff CEX
To establish construct validity of the Handoff Mini-CEX, we adapted a protocol used by Holmboe and colleagues to report the construct validity of the Handoff Mini-CEX, which is based on development and use of video scenarios depicting varying levels of clinical performance. (14) A clinical scenario script, based on prior observational work, was developed which represented an internal medicine resident (the sender) signing out three different patients to colleagues (intern (PGY-1) and resident). This scenario was developed to explicitly include observable components of professionalism, communication and setting. Three levels of performance – superior, satisfactory and unsatisfactory – were defined and described for the three domains. These levels were defined and then separate scripts were written using this information, demonstrating varying levels of performance in each of the domains of interest, using the descriptive anchors of the Handoff Mini-CEX. (See Figure 2) After constructing the superior, or gold standard, script which showcases superior communication, professionalism and setting, individual domains of performance were changed (e.g. to satisfactory or unsatisfactory), while holding the other two constant at the superior level of performance. For example, superior communication requires that the sender provides anticipatory guidance and includes clinical rationale while unsatisfactory communication includes vague language about overnight events and a disorganized presentation of patients. Superior professionalism requires no inappropriate comments by the sender about patients, family, and staff as well as a presentation focused on the most urgent patients. Unsatisfactory professionalism is shown by a hurried and inattentive sign-out with inappropriate comments about patients, family, and staff. Finally, a superior setting is one in which the receiver is listening attentively and discourages interruptions while an unsatisfactory setting finds the sender or receiver answering pages during the hand-off and surrounded by background noise. We omitted the satisfactory level for “setting” due to the difficulties in creating subtleties in the environment. Permutations of each of these domains resulted in six scripts depicting different levels of sender performance (see Appendix 3). Only the performance level of the sender was changed and the receivers of the hand-off performance remained consistent and using best practices for receivers, such as attentive listening, asking questions, reading- back, and taking notes during the hand-off. The scripts were developed by two investigators (VA, SB) then reviewed and edited independently by other investigators (JF, PS) to achieve consensus. Actors were recruited to perform the video scenarios and were trained by the physician-investigators (JF, VA). The part of the sender was played by one of the study investigators (PS) with prior acting experience, and who had accrued over 40 hours of experience observing handoffs to depict varying levels of hand-off performance. The digital video recordings ranged in length from 2.00 minutes to 4.08 minutes. All digital videos were recorded using Sony XDCAM PMW-EX3 HD Camcorder.
Figure 2. Faculty Ratings* of Professionalism by Performance.
*The Hand-off CEX ratings are a nine-point scale: 1–3 = unsatisfactory, 4–6 = satisfactory, 7–9 = superior
Participants
Faculty from the University of Chicago Medical Center and Yale University were included. At the University of Chicago, faculty were recruited to participate via email by the study investigators to the Research in Medical Education (RIME) listhost which includes program directors, clerkship directors, and medical educators. Two sessions were offered and administered. Continuing medical education (CME) credit was provided for participation as this workshop was given in conjunction with the RIME CME conference. Evaluations were de-identified using a unique identifier for each rater. At Yale University, the workshop was offered as part of 2 seminars for program directors and chief residents from all specialties on handoffs. During these seminars, program directors and chief residents used anonymous evaluation rating forms that did not capture rater identifiers. No other incentive was provided for participation. Although neither faculty at the University of Chicago nor Yale received any formal training on handoff evaluation, they did receive a short introduction to the importance of handoffs and the goals of the workshop. The protocol was deemed exempt by the Institutional Review Board at the University of Chicago.
Workshop Protocol
After a brief introduction, faculty viewed the tapes in random order on a projected screen. Participants were instructed to use the Handoff Mini-CEX, to rate whichever element(s) of hand-off quality they believed they could suitably evaluate while watching the tapes. The videos were rated on the Handoff Mini-CEX, form and participants anonymously completed the forms independently without any contact with other participants. The lead investigators proctored all sessions. At University of Chicago, participants viewed and rated all six videos over the course of an hour. At Yale, due to time constraints in the program director and chief resident seminars, participants reviewed 1 of the videos in seminar 1 (unsatisfactory professionalism) and 2 in the other seminar (unsatisfactory communication, unsatisfactory professionalism – see Table 1).
Table 1.
Script Matrix
| Unsatisfactory | Satisfactory | Superior | |
|---|---|---|---|
| Communication |
Script 3 (n=36)*
“Look in the record; I’m sure it’s in there. And oh yeah, I need you to check enzymes and finish ruling her out” |
Script 2 (n=13)
“So the only thing to do is to check labs. You know, check CBC and cardiac enzymes.” |
Script 1 (n=13)
“So for today, I need you to check post-transfusion hemoglobin to make sure it’s back to the baseline of 10. If it’s under 10 then transfuse her 2 units but hopefully it will be bumped up. Also continue to check cardiac enzymes; the next set is coming at 2pm and we need to continue the rule out. If her enzymes are positive or she has other EKG changes, definitely call the cardio fellow, since they’ll want to take her to the CCU.” |
| Professionalism |
Script 5 (n=39)*
“…definitely call the cards fellow, since they’ll want to take her to the CCU. And let me tell you, if you don’t call her, she’ll rip you a new one.” |
Script 4 (n=22)*
“Let’s breeze through them quickly so I can get out of here, I’ve had a rough day. I’ll start with the sickest first, and oh my God she’s a train wreck!” |
Script 1
|
| Setting |
Script 6 (n=13)
|
Script 1
|
Denotes video scenario seen by Yale raters. All videos seen by University of Chicago raters
Data Collection and Statistical Analysis
Using combined data from University of Chicago and Yale University, descriptive statistics are reported on raw scores on the Handoff Mini-CEX. To assess internal consistency of the tool, Cronbach’s alpha was used. To assess inter-rater reliability of these attending physician ratings on the tool, we performed a Kendall’s coefficient of concordance after collapsing the ratings into 3 categories (unsatisfactory, satisfactory, superior). In addition, we also calculated intraclass correlation coefficients for each item using the raw data and generalizability analysis to calculate the number of raters that would be needed to achieve a desired reliability of 0.95. To ascertain if faculty were able to detect varying levels of performance depicted in the video, an ordinal test of trend on the communication, professionalism, and setting scores was performed.
To assess for rater bias, we were able to use the identifiers on the University of Chicago data to perform a two-way analysis of variance (ANOVA) to assess if faculty’s scores were associated with performance level after controlling for faculty. The results of the faculty rater coefficients and p values in the two-way ANOVA were also examined for any evidence of rater bias. All calculations were performed in STATA 11.0 (College Station, TX) with statistical significance defined as p<0.05.
Results
Forty-seven (47) faculty members (14=site 1; 33=site 2) participated in the validation workshops (two at UC, and two at Yale) which were held in August and September of 2011 providing a total of 172 observations of a possible 191 (90%).
The overall hand-off quality ratings for the superior, gold standard video (superior communication, professionalism and communication) ranged from 7–9 with a mean of 8.5(SD 0.7). The overall ratings for the video depicting satisfactory communication (satisfactory communication, superior professionalism and setting) ranged from 5–9 with a mean of 7.3(SD 1.1). The overall ratings for the unsatisfactory communication (unsatisfactory communication, superior professionalism and setting) video ranged from 1–7 with a mean of 2.6 (SD 1.2). The overall ratings for the satisfactory professionalism video (satisfactory professionalism, superior communication and setting) ranged from 4–8 with a mean of 5.7 (SD 1.3). The overall ratings for the unsatisfactory professionalism (unsatisfactory professionalism, superior communication and setting) video ranged from 2–5 with a mean of 2.4 (SD 1.03). Finally, the overall ratings for the unsatisfactory setting (unsatisfactory setting, superior communication and professionalism) video ranged from 1–8 with a mean of 3.1 (SD 1.7).
Figure 3 demonstrates that for the domain of communication, the raters were able to discern the unsatisfactory performance but had difficulty reliably distinguishing between superior and satisfactory performance. Figure 4 illustrates that for the domain of professionalism, raters were able to detect the videos’ changing levels of performance at the extremes of behavior, with unsatisfactory and superior displays more readily identified. Figure 5 shows that for the domain of setting, the raters were able to discern the unsatisfactory versus superior level of the changing setting. Of note, we also found a moderate significant correlation between ratings of professionalism and communication (r=0.47, p<0.001).
Figure 3. Faculty Ratings* of Setting by Performance.
*The Hand-off CEX ratings are a nine-point scale: 1–3 = unsatisfactory, 4–6 = satisfactory, 7–9 = superior
The Cronbach’s alpha, or measurement of internal reliability and consistency, for the Handoff Mini-CEX, (3-items plus overall) was 0.77, indicating high internal reliability and consistency. Using data from University of Chicago where raters were labeled with a unique identifier, the Kendall’s coefficient of concordance was calculated to be between 0.79, demonstrating high, inter-rater reliability of the faculty raters. High inter-rater reliability was also seen using intraclass coefficients for each domain: communication (0.84), professionalism (0.68), setting (0.83) and overall (0.89). Using generalizability analysis, the average reliability was determined to be above 0.9 for all domains (0.99 for overall).
Lastly, the two-way ANOVA (n=75 observations from 13 raters) revealed no evidence of rater bias when examining the coefficient for attending rater (p=0.55 for professionalism;, p=0.45 for communication;, p=0.92 for setting). The range of scores for each video, however, was broad. (See Table 2)
Table 2.
Faculty’s Mini-Hand-off CEX Ratings by Level of Performance Depicted in Video*
| Unsatisfactory | Satisfactory | Superior | ||||||||
|---|---|---|---|---|---|---|---|---|---|---|
|
| ||||||||||
| Mean | Median | Range | Mean | Median | Range | Mean | Median | Range | P† | |
| Professionalism | 2.3 | 2 | 1–4 | 4.4 | 4 | 3–8 | 7.0 | 7 | 3–9 | 0.026 |
| Communication | 2.8 | 3 | 1–6 | 7 | 8 | 5–9 | 6.6 | 7 | 1–9 | 0.005 |
| Setting | 3.1 | 3 | 1–8 | 7.5 | 8 | 2–9 | 0.005 | |||
CEX ratings are on a nine-point scale:1–3=unsatisfactory, 4–6=satisfactory, 7–9=superior
p value from two-way ANOVA examining the level of performance on rating of that construct controlling for rater
Discussion
This study demonstrates that valid conclusions on handoff performance can be drawn using the Handoff CEX as the instrument to rate handoff quality. Utilizing standardized videos depicting varying levels of performance communication, professionalism, and setting, the Handoff Mini-CEX has demonstrated potential to discern between increasing levels of performance, providing evidence for the construct validity of the instrument.
We observed that faculty could reliably detect unsatisfactory professionalism with ease and that there was a distinct correlation between faculty ratings and the internally set levels of performance displayed in the videos. This trend demonstrated that faculty were able to discern different levels of professionalism using the Handoff Mini-CEX. It became more difficult, however, for faculty to detect superior professionalism when the domain of communication was permuted. If the sender of the hand-off was professional but the information delivered was disorganized, inaccurate and missing crucial pieces of information, the faculty perceived this ineffective communication as unprofessional. Prior literature on professionalism has found that communication is a necessary component of professional behavior and consequently being a competent communicator is necessary to fulfill ones duty as a professional physician.(15, 16) This is of note since we did find a moderate significant correlation between ratings of professionalism and communication. It is possible that this distinction would be made clearer with formal rater training in the future prior to any evaluations. However, it is also possible that it is not entirely possible to separate professionalism and communication due to a synergistic role between the two domains. If this is the case, it would be important to educate clinicians to present patients in a concise, clear and accurate way with a professional demeanor. Acknowledging professional responsibility as an integral piece of patient care is also critical in effectively communicating patient information.(5)
We also noted that faculty could detect unsatisfactory communication consistently; however, they were unable to differentiate between satisfactory and superior communication reliably or consistently. Since the unsatisfactory professionalism, unsatisfactory setting and satisfactory professionalism videos all demonstrated superior communication; we believe that the faculty penalized communication when distractions, in the form of interruptions and rude behavior by the resident giving the hand-off, interrupted the flow of the hand-off. Thus, the wide ranges in scores observed by some raters may be attributed to this interaction between the Handoff Mini-CEX domains. In the future, definitions of the anchors, including at the middle spectrum of performance, and rater training may improve the ability of raters to distinguish performance between each domain.
The overall value of the Handoff mini-CEX is in its ease of use, in part due to its brevity, as well as evidence for its validity in distinguishing between varying levels of performance. Given the emphasis on monitoring handoff quality and performance, the Handoff Mini-CEX provides a standard foundation from which baseline handoff performance can be easily measured and improved. Moreover, it can also be used to give individual feedback to a specific practicing clinician on their practices and an opportunity to improve. This is particularly important given current recommendations by the Joint Commission that handoffs are standardized and by the ACGME that residents are competent in handoff skills. Moreover, given the creation of SHM’s handoff recommendations and handoffs as a core competency for hospitalists, the tool provides the ability for hospitalist programs to actually assess their handoff practices as baseline measurements for any quality improvement activities that may take place.
Faculty were able to discern the superior and unsatisfactory levels of setting with ease. After watching and rating the videos, participants said that the chaotic scene of the unsatisfactory setting video had significant authenticity and that they were constantly interrupted during their own hand-offs by pages, phone calls and people entering the hand-off space. System-level fixes such as protected time and dedicated space for hand-offs and discouraging pages to be sent during the designated hand-off time could mitigate the reality of unsatisfactory settings. (17, 18)
Our study has several limitations. First, although this study was held at a two sites, it included a small number of faculty that can impact the generalizability of our findings. Implementation varied at Yale and University of Chicago, preventing use of all data for all analyses. Furthermore, institutional culture may also impact faculty raters’ perceptions, so future work aims at repeating our protocol at partner institutions, increasing both the number and diversity of participants. We are also unable to compare the new shorter Handoff Mini-CEX to the larger 9-item Handoff CEX in this study.
Despite these limitations, we believe that the Handoff Mini-CEX, has future potential as an instrument with which to make valid and reliable conclusions about hand-off quality and could be used to both evaluate hand-off quality and as an educational tool for trainees and faculty on effective hand-off communication.
Supplementary Material
Figure 1. Faculty Ratings* of Communication by Performance.
*The Hand-off CEX ratings are a nine-point scale: 1–3 = unsatisfactory, 4–6 = satisfactory, 7–9 = superior
Acknowledgments
Funding: National Institute on Aging Short-Term Aging-Related Research Program (5T35AG029795), Agency for Healthcare Research & Quality (1 R03HS018278-01), and the University of Chicago Department of Medicine Excellence in Medical Education Award. Dr. Horwitz is supported by the National Institute on Aging (K08 AG038336) and by the American Federation for Aging Research through the Paul B. Beeson Career Development Award Program. Dr. Arora is funded by National Institute on Aging Career Development Award K23AG033763.
Footnotes
Prior presentations of this data include the Association of American Medical Colleges 2011 Meeting in Denver, CO, the 2012 Association of Program Directors of Internal Medicine Meeting in Atlanta, GA, and the Society of General Internal Medicine Meeting in Orlando, FL in 2012.
References
- 1.Nasca TJ, Day SH, Amis ES. The new recommendations on duty hours from the ACGME task force. New Engl J Med. 2010;363(2):e3. doi: 10.1056/NEJMsb1005800. [DOI] [PubMed] [Google Scholar]
- 2. [Accessed on February 8, 2014];ACGME Common Program Requirements Effective. 2011 Jul 1; Available at: http://www.acgme.org/acgmeweb/Portals/0/PDFs/Common_Program_Requirements_07012011[2].pdf.
- 3.Horwitz LI, Moin T, Krumholz HM, Wang L, Bradley EH. Consequences of inadequate sign-out for patient care. Archives of Internal Medicine. 2008;168(16):1755–60. doi: 10.1001/archinte.168.16.1755. [DOI] [PubMed] [Google Scholar]
- 4.Arora V, Johnson J, Lovinger D, Humphrey HJ, Meltzer DO. Communication failures in patient sign-out and suggestions for improvement: A critical incident analysis. Qual Saf Healthcare. 2005;14(6):401–7. doi: 10.1136/qshc.2005.015107. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Arora VM, Manjarrez E, Dressler DD, Basaviah P, Halasyamani L, Kripalani S. Hospitalist hand-offs: A systematic review and task force recommendations. J Hosp Med. 2009;4(7):433–40. doi: 10.1002/jhm.573. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Joint Commission on Accreditation of Healthcare Organizations. Joint Commission announces 2006 national patient safety goals for ambulatory care and office-based surgery organizations. 2005 Sep 1; [Internet] Available from: http://www.jcaho.org/news+room/news+release+archives/06_npsg_amb_obs.htm. [PubMed]
- 7.World Health Organization Collaborating Centre for Patient Safety Solutions on Communication During Patient Hand-Overs. 2007;1 solution 1. Available from: http://www.who.int/patientsafety/solutions/patientsafety/PS-Solution3.pdf. [Google Scholar]
- 8.Patterson ES, Wears RL. Patient handoffs: standardized and reliable measurement tools remain elusive. Jt Comm J Qual Patient Saf. 2010;36(2):52–61. doi: 10.1016/s1553-7250(10)36011-9. [DOI] [PubMed] [Google Scholar]
- 9.Horwitz L, Rand D, Staisiunas P, Van Ness P, Araujo K, Farnan JM, Arora VM. Development of a handoff evaluation tool for shift-to-shift physician handoffs: The handoff CEX. J Hosp Med. 2013;8(4):191–200. doi: 10.1002/jhm.2023. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Farnan JM, Paro JAM, Rodriguez RM, Reddy ST, Horwitz LI, Johnson JK, et al. Hand-off education and evaluation: Piloting the observed simulated hand-off experience (OSHE) J Gen Intern Med. 2010;25(2):129–34. doi: 10.1007/s11606-009-1170-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Horwitz LI, Dombroski J, Murphy TE, Farnan JM, Johnson JK, Arora VM. Validation of a Handoff Tool: the Handoff CEX. J Clin Nurs. 2013;22(9–10):1477–86. doi: 10.1111/j.1365-2702.2012.04131.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Norcini JJ, Blank LL, Duffy FD, Fortna GS. The mini-CEX: A method for assessing clinical skills. Ann Intern Med. 2003;138(6):476–81. doi: 10.7326/0003-4819-138-6-200303180-00012. [DOI] [PubMed] [Google Scholar]
- 13.Patterson ES, Roth EM, Woods DD, Chow R, Gomes JO. Handoff strategies in settings with high consequences for failure: lessons for health care operations. Int J Qual Health Care. 2004;16(2):125–32. doi: 10.1093/intqhc/mzh026. [DOI] [PubMed] [Google Scholar]
- 14.Holmboe ES, Huot S, Chung J, Norcini J, Hawkins RE. Construct validity of the miniclinical evaluation exercise (miniCEX) Acad Med. 2003;78(8):826–30. doi: 10.1097/00001888-200308000-00018. [DOI] [PubMed] [Google Scholar]
- 15.Reddy ST, Farnan JM, Yoon JD, Leo T, Upadhyay GA, Humphrey HJ, et al. Third-year medical students’ participation in and perceptions of unprofessional behaviors. Acad Med. 2007;82(10 Suppl):S35–9. doi: 10.1097/ACM.0b013e3181405e1c. [DOI] [PubMed] [Google Scholar]
- 16.Hafferty FW. Professionalism--the next wave. N Engl J Med. 2006;355(20):2151–2. doi: 10.1056/NEJMe068217. [DOI] [PubMed] [Google Scholar]
- 17.Chang VY, Arora VM, Lev-Ari S, D’Arcy M, Keysar B. Interns overestimate the effectiveness of their hand-off communication. Pediatrics. 2010;125(3):491–6. doi: 10.1542/peds.2009-0351. [DOI] [PubMed] [Google Scholar]
- 18.Greenstein EA, Arora VM, Staisiunas PG, Banerjee SS, Farnan JM. Characterising physician listening behaviour during hospitalist handoffs using the HEAR checklist. BMJ Qual Saf. 2013;22(3):203–9. doi: 10.1136/bmjqs-2012-001138. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.



