Abstract
Background:
Many orthopedic surgeries involve the challenging integration of fluoroscopic image interpretation with skillful tool manipulation to enable procedures to be performed through less invasive approaches. Simulation has proved beneficial for teaching and improving these skills for residents, but similar benefits have not yet been realized for practicing orthopedic surgeons. A vision is presented to elevate community orthopedic practice and improve patient safety by advancing the use of simulators for training and assessing surgical skills.
Methods:
Key elements of this vision that are established include 1) methods for the objective and rigorous assessment of the performance of practicing surgeons now exist, 2) simulators are sufficiently mature and sophisticated that practicing surgeons will use them, and 3) practicing surgeons can improve their performance with appropriate feedback and coaching.
Results:
Data presented indicate that surgical performance can be adequately and comparably measured using structured observations made by experts or non-expert crowds, with the crowdsourcing approach being more expedient and less expensive. Rigorous measures of the surgical result and intermediate objectives obtained semi-automatically from intra-operative fluoroscopic image sequences can distinguish performances of experts from novices. Experience suggests that practicing orthopedic surgeons are open to and can be constructively engaged by a family of mature simulators as a means to evaluate and improve their surgical skills.
Conclusions:
The results presented support our contention that new objective assessment measures are sufficient for evaluating the performance of working surgeons. The novel class of orthopedic surgical simulators available were tested and approved by practicing physicians. There exists a clear opportunity to combine purpose-designed simulator exercises with virtual coaching to help practicing physicians retain, retrain, and improve their technical skills. This will ultimately reduce cost, increase the quality of care, and decrease complication rates.
Clinical Relevance:
This vision articulates a means to boost the confidence of practitioners and ease their anxiety so that they perform impactful procedures more often in community hospitals, which promises to improve treatment and reduce the cost of care while keeping patients closer to their homes and families.
Keywords: surgical skills training, performance assessment
Introduction
The manner in which orthopedic surgical skills are trained and assessed is rapidly changing. New training methods, higher quality simulators, public health concerns, and a mandate from certifying bodies have together catalyzed the structured teaching of surgical skills outside of the operating room (OR) across surgical specialties.1-3 Unfortunately, the ubiquitous Halstedian model of training is stymied by limited resources for the evaluation of technical competency. Worse, existing evaluation methods may be flawed; suffering from bias, impractical for high-volume use, and of unverified accuracy. Programs routinely credential surgeons for procedures they have never done.4 Self-reported caseloads are inconsistent across institutions, and caseload or academic rank provide imperfect measures of true technical skill.5,6
Compared to residency training, the adoption of simulation in the larger community of practicing surgeons has lagged behind. This is an important missed opportunity. Surgeons with lower technical skills demonstrate a 5x higher mortality rate, 3x higher complication rate, and 29% longer operative time compared to top-quartile technical skills performers.7 Indeed, with over 135,000 surgeons performing 51,000,000 procedures in the U.S. each year,8 identifying surgeons at high risk for poor patient outcomes and providing them with targeted coaching is an acute, high-priority need.
Community practitioners could benefit from targeted training. In orthopedics, common fracture surgeries, such as the operative reduction of articular fractures and pinning of pediatric elbow fractures, are too often referred from community practitioners to tertiary care centers or treated using out-of-date techniques.9,10 In fact, the skill required to pin a pediatric elbow fracture is among those with the lowest self-reported competence among orthopedic surgeons, despite elbow fractures accounting for approximately 8% of all pediatric fractures.11 This presents an opportunity for simulation and coaching to improve the community practice of orthopedic trauma surgery. At the nexus of this opportunity is the skillful use of fluoroscopy, which enables surgical procedures to be performed through less invasive approaches. Less invasive approaches reduce the risk of wound complications, lead to quicker recovery times, and improve patient comfort, all while ensuring comparable results, at least for experts.12
Unfortunately, not all orthopedic surgeons are adept with the skills that enable these less invasive approaches, and others lack the confidence to utilize them. Two target surgeries are emblematic: articular fractures of the distal tibia and pediatric supracondylar humerus fractures. Complex articular fractures require precise reduction to achieve optimal outcomes. These surgeries have historically been done through extensile open approaches that risk further trauma to compromised soft tissues. More limited approaches are associated with fewer complications, but they rely heavily on fluoroscopic guidance and skillful reduction maneuvers. Similarly, when treating a pediatric supracondylar humerus fracture, a surgeon aims to restore pre-injury position and alignment. Then K-wires are placed across the fracture site to help maintain alignment until the fracture heals, while protecting the growth plate in these young children. Supracondylar humerus fractures are the most common pediatric fracture treated surgically worldwide.13 Their treatment requires skill that takes time and concerted effort to master.14 Many community surgeons do not feel competent to perform this surgery.15 This leads to referrals that delay care, and it also costs more to be treated at a tertiary care facility than at a community hospital.16
We propose that the coupling of simulation with virtual coaching provides a unique opportunity to facilitate retaining, retraining, and improving the technical skills of practicing surgeons, which can reduce the cost of care, increase its quality, and decrease complications. In the following we present the rationale and arguments for the potential of this approach and a path forward. This vision to improve the community practice of orthopedic trauma surgery builds upon three recent fundamental advancements. First, unique and rigorous analysis approaches for modeling and assessing surgical skill based upon fluoroscopic sequences and/or video collected in the OR have been developed.17,18 Second, simulators that incorporate technologies mixing physical realism with synthetic fluoroscopy images19-21 have been shown to be effective,22 and they capture exciting and attractive technical targets for practicing surgeons. Finally, the simulators and assessment procedures facilitate virtual coaching that leverages scientifically-supported formative and summative feedback approaches imbued with a deep understanding of successful surgical behaviors.
Methods
Assessment Techniques are Ready to Measure the Skill Level of Practicing Surgeons
Most technical skills assessment methods employ video-based evaluation by expert surgeons utilizing survey instruments based on the Objective Structured Assessment of Technical Skills (OSATS).23 In bariatric surgery, OSATS scores have been shown to correlate with patient outcomes,7 but this correlation has not been well studied in orthopedics. Structured survey instruments such as OSATS23 are subject to confirmation bias in expert ratings, and OSATS scores fail to correlate with surgical results measured in a laboratory setting.24,25
A key challenge to establishing simulator skill transfer to the OR is finding a way to objectively measure OR performance. We have pioneered a technique analyzing surgical behavior from fluroscopic image sequences. Our experience has shown that important insights can be extracted from a surgical fluoroscopic image sequence,18,20,26 including details of surgical strategy, surgical skill, and the quantitative assessment of the final implant placement. One challenge in obtaining these insights involves collecting detailed measurements from as many as 150 images in a sequence for a long or complex surgery.
To address this challenge, we developed a semi-autonomous process involving artificial intelligence and customized image analysis software. The software speeds the sequence analysis so that dozens of surgical images can be analyzed by a non-surgeon in minutes, rather than days. The original implementation was tailored toward analyzing fluoroscopic sequences of a guide wire being placed during intertrochanteric hip fracture reduction and fixation. The analysis provides the wire entry-point, the actual wire trajectory versus that desired, the duration of the surgical step, the number of images collected, and the tip-apex distance (TAD). A large TAD is a strong predictor of later implant failure.27,28 For this reason, a central surgical objective is to minimize the TAD, but it can come at the cost of substantial fluoroscopic radiation and additional operative time when skills are poorly developed. The details extracted from these OR images has proven useful in distinguishing expert from novice performance.
A more recent adaptation of the software analyzes the pinning of pediatric supracondylar humerus fractures (Figure 1).29 We are extending the software to analyze the reduction and fixation of distal tibia fractures, as well. This semi-automated analysis approach has yielded powerful quantitative, repeatable insight into surgeon behavior that allows us to assess skill in new ways. The technique is also uniquely suited to compare performance in simulation and in the OR. Finally, the technique provides insights that lead directly to timely formative feedback, because analysis at the decision-making level captures each step of a surgical procedure.
Figure 1.

Automatically detected K-wires during the pinning of a pediatric supracondylar humerus fracture. Calculated performance metrics include the angles of the wires relative to the fracture plane, the spacing of the wires at the fracture plane, and the wire spread ratio.
Another problem with OSATS video evaluation methods is that they are impractical for widespread use, largely due to the excessive time burden demanded of experts or faculty raters. This motivated the pioneering use of “crowds” of non-expert raters to provide inexpensive evaluations that proved statistically concordant with expert ratings,30 obtaining thousands of ratings at unprecedented speeds.31,32 This method has come to be called crowd-sourced assessment of technical skill (CSATS). Unlike expert raters, ratings aggregated from non-expert crowds are fast (available within hours of video upload), scale to practically unlimited numbers of ratings, and are available for repeated ratings around the clock at near-minimum wage. Further, CSATS ratings prove statistically favorable compared to typical panels of experts or faculty and correlate well to patient outcomes.33
Simulator Technologies Are Ready for Practicing Surgeons
The key to developing an effective simulator is understanding and defining specific educational objectives and the manner in which performance is measured. Properly used, simulators can be more effective than traditional training approaches.34 They are also useful for identifying skill performance issues.35 Thus, simulation makes an attractive target for assessing, training, and retraining the skills of practicing surgeons.
Our group has developed and tested simulators for navigating surgical wires using fluoroscopic guidance and for reducing fractures through limited open incisions, two common and difficult skills in orthopedic surgery. These simulators feature replaceable, plastic foam bones and synthetic fluoroscopic imagery to help surgeons recognize and interpret key image features, connecting their hand movements to surgical goals. Our wire navigation orthopedic surgical simulation platform replicates the look and feel of navigating a wire through bone using fluoroscopic guidance (Figure 2).19 Two differentiating features of the platform are: (1) camera-based tracking of the wire replaces fluoroscopic radiation exposure, and (2) the foam bone surrogate replicates the feel of drilling through actual bone.
Figure 2.

The orthopedic surgical simulation platform relies upon camera-based tracking of a laser-etched stainless steel K-wire, otherwise identical to that used in surgery. In the case of the hip wire navigation simulation shown, a foam bone surrogate is housed within a soft tissue mimicking sleeve to obscure the trainee’s view of the object.
We initially developed this platform to simulate the navigation of a wire in the treatment of intertrochanteric hip fractures. After several years of development and testing, this hip fracture wire navigation simulator19,25 is currently under study by the American Board of Orthopaedic Surgery (ABOS) and is used twice annually as part of resident fracture courses offered by the Orthopaedic Trauma Association (OTA). Seeking to get more data from advanced users, we tested at the OTA Trauma Fellows course in 2019, observing the performance of 30 fellows near the pinnacle of their surgical training in orthopedic trauma. We have demonstrated that the platform is extensible to other relevant orthopedic applications,20 including the pinning of a pediatric supracondylar humerus fracture (Figure 3).36
Figure 3.

The pediatric supracondylar humerus fracture simulator provides residents with the opportunity to place three wires across a fracture. The position of the wires on the plastic bone (left) are represented in synthetic fluoroscopic images rendered by the simulator (right). Performance is measured by the spread of the wires across the fracture line, the time taken in placing the wires, and the number of fluoroscopic images obtained along the way.
Our team has also recently developed an articular fracture reduction simulator to help train more limited surgical approaches that are associated with faster recovery and fewer complications (Figure 4).21,37 This is accomplished through an electromagnetic system that provides real-time tracking of surgical tools and artificial fracture fragments buried in a soft tissue model. The system provides quantitative feedback on 17 different performance metrics including novel features, such as a summative snapshot of fracture displacement over time.
Figure 4.

(a) Articular fracture reduction simulator. (b) The articular fragment is tethered to resist reduction to mimic the clinical setting. (c) Representation of tenaculum holding reduction as a K-wire is inserted as fixation.
Although each of these simulators was initially developed for training residents, we have begun to use them to assess experienced surgeons, typically to provide a comparative performance assessment or training goal. These experiences with practicing surgeons suggest that the simulators are capable of distinguishing skill levels among practicing surgeons and that surgeons find the experience challenging and enjoyable, as described in the results section.
Simulator Training is Likely to Improve Performance
Constructive feedback strategies are essential for effective skill training and performance improvement,38,39 but different types of feedback are more or less effective at different stages in skill mastery. Formative feedback is information provided to the learner during an educational exercise to modify behavior and improve learning.40 In general, timely detailed formative feedback is most effective for early training stages, and delayed, summative feedback is more effective for later retaining, retraining, and improving performance. Experienced surgeons may find immediate, detailed, formative feedback frustrating. For instance, the formative feedback currently provided with the wire navigation simulator includes recommendations such as “drop your hand” or “shift the wire angle 5° distally.” This level of detail, effective for novice residents, projects a presumed procedural approach for which an experienced surgeon may have developed their own effective and appropriate alternative strategies.
Summative feedback, often more effective for advanced learners, provides information about the results of an entire procedure that is generally intended to provide the basis of comparison with peers or an external standard.41 Our articular fracture reduction simulator currently emphasizes summative feedback on the duration and the precision of the fracture reduction, for example. Summative feedback is typically provided upon completion of the procedure, but intermediate extrapolation can be used to project some elements of performance, like reduction quality. This type of feedback can be useful for someone who is fine-tuning details of their approach as they integrate the tradeoffs between different steps of a procedure. However, if the learner is unclear on appropriate strategies to achieve particular intermediate goals, summative feedback can be less effective than formative feedback.
There are many ways to provide feedback and not all are equally helpful. For example, providing an explanation for why an answer was wrong is 50% more effective than providing the right answer and 10 times more effective than saying that a given answer was incorrect.42 The Interactive Tutor Model is one approach to understanding the diversity of factors that affect feedback design in order to produce effective feedback.43 This model depends on a close match between the learner and the instructors’ understanding of the objectives and intermediate goals. With a loosely defined surgical task, different surgeons may approach the task with different strategies, all equally acceptable. If the feedback presumes one goal when the surgeon is appropriately choosing another, it may decrease the surgeon’s trust in the automated feedback, weakening its effectiveness.44
Results
The following sections summarize data from various experiments that we have conducted over the past decade supporting the three foundational arguments that underlie the vision of expanding simulation training to practicing surgeons.
Assessment Techniques Can Measure Transfer of Training
The predominant assessment techniques have important flaws. Surgical assessment approaches have assumed that faculty ratings are the most reliable measure of performance and identity-blind video ratings are the “gold-standard” for assessment. But mounting evidence suggests that both approaches are flawed. Experts appear to reward speed over technical accuracy.6 Preliminary results suggest that expert raters’ expectations of a surgeon’s technical skill based on expertise/reputation do not agree with their own scores assigned based upon anonymized video.45,46 When two faculty coders reviewed 56 anonymized videos of benchtop laparoscopic exam tasks, they identified 30 as below average competency or non-competent. They were not told that these were videos of surgeons whom they had just nominated as “experts” by reputation or academic rank (Figure 5a).46,47
Figure 5.

(a) A panel of two faculty surgeons disqualified over 50% of videos from surgeons reputed to be experts when blinded to their identity. [Figure reprinted with permission from.45] (b) Concordance between experts and crowds for three different studies. [Figure reprinted with permission from.48] (c) On average, expert faculty spent 6 minutes reviewing a 13-minute video, while non-expert crowd workers from Amazon Mechanical Turk CSATS spent 16 minutes reviewing the video33.
One lesson we learned from this experience is how onerous it can be to have faculty surgeons grade 56 videos. This motivated seminal work in 2011 to test whether ratings from non-expert crowds are concordant with expert ratings. It was found that when sufficient non-expert crowd workers (~35) from Amazon Mechanical Turk used a survey to assess video, their aggregate scores agreed remarkably well with expert ratings (Figure 5b). These results were confirmed for laparoscopy,32 robotic surgery,30,49 animate training,50 and real patients.33 The crowd-sourced ratings even correlated well with patient outcomes.33,51 The crowd-sourcing approach provided 16,418 ratings for 430 videos at 1.3 ratings/min for under $5000.52 Finally, deeper inspection revealed that expert raters do not watch full videos (Figure 5c), and in some cases, non-expert ratings appear to predict surgical outcomes better than experts. For these reasons, the CSATS method enjoys increasing adoption in medical skills assessment research.30,31,49,50,53-56 The crowd-sourcing approach provides an inexpensive and expeditious method for analyzing surgical performance both with simulators and in the OR. Another objective assessment method links performance to specific surgical behaviors.
We evaluated the transfer of training on the wire navigation simulator in both a mock OR22 and in the actual OR (Figure 6).17,18 This allowed us to begin to differentiate performance improvements caused by simulator training from those that accrue with experience. We define consistent measures that are common to both the simulator and the actual OR. For example, residents were assessed in the mock OR based on their use of fluoroscopy, total time, and TAD. Residents that trained on the simulator had a lower TAD than those who did not (p = 0.001), and their performance on the simulator (i.e., TAD, image use, and overall time) was correlated with performance in the mock OR.22 This assessment approach was clearly sufficient to measure the performance gain achieved by residents with some focused training, but as the next section will demonstrate, it was also sufficient to distinguish different cohorts of experts.
Figure 6.

The current orthopedic wire navigation surgical simulator was validated by measuring later performance in a mock OR (left). Performance measures were defined from information that can also be collected in the actual OR (right) to test skill transfer.
Simulators are Ready for Practicing Surgeons
In addition to the numerous tests with first- and second-year residents, in November 2019 we attended the Hennepin Orthopaedic and Trauma Seminar in Minneapolis to gather data on the performance of community orthopedic surgeons. After orientation to the simulator, 17 community surgeons completed three assessments with the hip fracture wire navigation. We compared the practicing surgeons’ performance with that of a cohort of residents we previously studied.22 Figure 7 illustrates normalized composite performance scores on the wire navigation. In this admittedly limited sample of community surgeons, we found their performance to be on average better than the residents, but there was a broad distribution of performance amongst both residents and community surgeons. Subjective rating questionnaires and informal interviews with the practicing surgeons confirmed that they felt the simulation was interesting, sufficiently realistic and challenging for practicing surgeons.
Figure 7.

A plot of the normalized composite scores of two cohorts of surgeons placing a hip wire with the wire navigation simulator. The data suggest that most community surgeons perform better than most residents, but that there are outliers in both groups that might benefit from additional training.
Results with the fracture reduction simulation promise even greater interest among practicing surgeons. Figure 8 presents performance results comparing an expert and a novice reducing a distal tibia fracture on the fracture reduction simulation. Although both surgeons achieved an acceptable final reduction (<2mm),57 the unique in-the-moment summative feedback provided by the simulator (Figure 8a) clearly demonstrates how the novice struggled in achieving their result, unnecessarily manipulating the fragment by moving it cumulatively 158 mm more than the expert. This line of inquiry also indicates that experienced surgeons utilize tools more efficiently during fracture reduction (Figure 8b).37 The novice surgeon lacked an understanding of proper tool placement to assist in the reduction, instead relying on an iterative method where they continuously attempted to replace their tool in equally ineffective locations, in all contacting the bone over a 517% larger area than the expert. As experts have attested when using the simulator, the unique, performance-specific feedback from the simulator is particularly helpful, because it provides viewpoints and measures that cannot be replicated in the OR environment. This provides practicing surgeons with the opportunity to reflect on their performance with a new perspective, offering new insights.
Figure 8.

(a) Fracture reduction accuracy (step-off) over time (seconds). Vertical red lines indicate the reduction period of a mock surgery. Compared to an expert (left), the novice (right) showed a large amount of fragment manipulation. Note the time scaling is compressed in the novice performance to fit. (b) Heat maps of tool and fragment interaction points. Blue = no interaction, yellow = an interaction lasting longer than 1.5 seconds. The tool-fragment interaction of the expert (left) differed notably from that of a novice (right).
Training is Likely to Improve the Skills of Practicing Surgeons
In the course of our work defining performance assessment metrics, we found limitations in the use of simple “final result” measures (e.g., the TAD) that do not account for in-the-moment decision-making that is a hallmark of performance. We came to recognize that it is precisely during these moments of what we have come to call micro-decisions, that there is a unique opportunity to teach. In response, we surveyed performance as reflected by the series of fluoroscopic images, and defined a short list of five errors we routinely observed, such as when a novice adjusts the wire in the wrong direction, or over compensates for an alignment error, or switches from the antero-posterior to the lateral view inappropriately. These clearly indicate errors in judgment that can be readily detected, and each error has an unambiguous corrective behavior that can be taught.
For the application in the hip, we have defined a composite score that includes the TAD metric, the number of decision errors made, and the average angle of the wire movement errors. Each metric was first normalized based on the overall population. A higher composite score indicates better performance. A score that is equal to zero would indicate that it matches perfectly with the average performance across all subjects. Our studies with residents have supported the idea that relevant surgical experience correlates with performance on the simulator. This shows the value of considering the behavior at the micro-decision level and also strengthens our ability to distinguish expert from novice performance.
The advantage of this micro-decision analysis framework, as well as the other objective performance measures, is that they provide benchmarks against which a specific performance can be assessed. This benchmarking strategy allows us to provide specific, timely feedback to surgeons as they work or immediately afterwards. Subjective ratings by experts and crowd-sourced assessments do not always provide such direct feedback. When timely, specific feedback is provided, it naturally results in improved performance.
Discussion
The results presented support our contention that the current assessment methods are sufficient for measuring the performance of working surgeons. We demonstrated the advantages of two approaches: wire position analysis and crowd sourcing. Figure 7 demonstrates that the simulator assessment approach can distinguish skill levels between a cohort of residents and practicing surgeons. It is important to remember that reliably measuring the performance of a single individual is more difficult, however, than measuring group characteristics. We did not yet demonstrate how much testing is required to quantify an individual as having sufficient skill. Measuring performance with repeated exercises will enable such an assessment. Providing a variety of exercises will make the activity more enjoyable and the measurement more robust. Overcoming the vagaries of bias and the challenge of noisy measurements was an essential obstacle for objectively measuring performance level and performance gains, which are both essential for training practicing surgeons.
The novel class of orthopedic surgical simulators we have developed were tested and approved by practicing physicians. The successful adaptation of the hip wire navigation simulator to pediatric elbow and illiosacral screws indicates that the simulators can be adapted as needed to create training environments that are clinically relevant and intellectually demanding for practicing physicians. Novel summative performance assessment provided by the fracture reduction simulator promises to open new opportunities for training and assessment that is not possible in the OR nor with typical cadaver exercises. The informal feedback and enthusiasm of practicing physicians who work with the simulators suggests that the simulator technology is ready for the next step.
Finally, the micro-decision analysis and quantitative analysis approaches to feedback generation make excellent input for a virtual coach. The micro-decision analysis can provide important observations of specific moments that a practicing surgeon could use to further refine his or her performance. The analytic analysis of the whole performance provides an opportunity for practicing surgeons to gain insights into their performance that go beyond the more rigidly defined decision analysis. The feedback is ready for practicing surgeons.
Our research approach leverages existing relationships with a regional consortium of residency programs, the ABOS, and the OTA. We have begun to expand our reach to include statewide Orthopedic Societies in Iowa and Minnesota, as well as an annual orthopedic and trauma seminar run in Minneapolis. This will allow us to extend our experience with recruiting and training hundreds of residents to the relatively new challenge of working with community practitioners. The work promises to extend our insight into how expertise is manifested and can be measured in the OR, how that can be compared with behaviors observed during simulated surgeries, what type of feedback is effective in simulation, and how this knowledge can be integrated into a virtual coach to make surgery more effective.
Efforts are already underway to refine our existing simulators so that they effectively allow working surgeons to practice their skills in a manner that will enable them to retain, retrain, or improve their performance. More generally, the simulators will improve skill in the broader area of navigating surgical wires with fluoroscopic guidance. The work will permit surgeons to rehearse these technical skills before further exercising them on live patients in the high-stakes atmosphere of the OR. The methods derived will boost the confidence of practitioners and ease their anxiety with performing these procedures in community hospitals, which will improve treatment and reduce the cost of care while keeping patients closer to their homes and families. We expect to demonstrate the clinical benefits of the surgical simulator and virtual coaching by increasing the number of procedures done by practitioners in their community hospital settings.
References
- 1.Ferguson PC, Kraemer W, Nousiainen M, Safir O, Sonnadara R, Alman B, Reznick R. Three-year experience with an innovative, modular competency-based curriculum for orthopaedic training. J Bone Joint Surg Am. 2013;95(21):e166. doi: 10.2106/JBJS.M.00314. [DOI] [PubMed] [Google Scholar]
- 2.Gallagher AG. Metric-based simulation training to proficiency in medical education: What it is and how to do it. Ulster Med J. 2012;81(3):107–13. [PMC free article] [PubMed] [Google Scholar]
- 3.Makary MA, Daniel M. Medical error-the third leading cause of death in the US. BMJ. 2016;353:i2139. doi: 10.1136/bmj.i2139. [DOI] [PubMed] [Google Scholar]
- 4.Bell RH, Biester TW, Tabuenca A, Rhodes RS, Cofer JB, Britt LD, Lewis FR. Operative experience of residents in US general surgery programs: a gap between expectation and experience. Ann Surg. 2009;249(5):719–24. doi: 10.1097/SLA.0b013e3181a38e59. [DOI] [PubMed] [Google Scholar]
- 5.Feldman LS, Hagarty SE, Ghitulescu G, Stan-bridge D, Fried GM. Relationship between objective assessment of technical skills and subjective in-training evaluations in surgical residents. J Am Coll Surg. 2004;198(1):105–10. doi: 10.1016/j.jamcollsurg.2003.08.020. [DOI] [PubMed] [Google Scholar]
- 6.Kowalewski TM, Sweet R, Lendvay TS, Menhadji A, Averch T, Box G, Brand T, Ferrandino M, Kaouk J, Knudsen B, Landman J, Lee B, Schwartz BF, McDougall E. Validation of the AUA BLUS tasks. J Urol. 2016;195(4 Pt 1):998–1005. doi: 10.1016/j.juro.2015.10.087. [DOI] [PubMed] [Google Scholar]
- 7.Birkmeyer JD, Finks JF, O’Reilly A, Oerline M, Carlin AM, Nunn AR, Dimick J, Banerjee M, Birkmeyer NJ, Michigan Bariatric Surgery C. Surgical skill and complication rates after bariatric surgery. N Engl J Med. 2013;369(15):1434–42. doi: 10.1056/NEJMsa1300625. [DOI] [PubMed] [Google Scholar]
- 8.Weiser TG, Regenbogen SE, Thompson KD, Haynes AB, Lipsitz SR, Berry WR, Gawande AA. An estimation of the global volume of surgery: a modelling strategy based on available data. Lancet. 2008;372(9633):139–44. doi: 10.1016/S0140-6736(08)60878-8. [DOI] [PubMed] [Google Scholar]
- 9.Kasser JR. Location of treatment of supracondylar fractures of the humerus in children. Clin Orthop Relat Res. 2005;434:110–3. doi: 10.1097/01.blo.0000162410.47731.69. [DOI] [PubMed] [Google Scholar]
- 10.Saarinen AJ, Helenius I. Paediatric supracondylar humeral fractures: the effect of the surgical specialty on the outcomes. J Child Orthop. 2019;13(1):40–6. doi: 10.1302/1863-2548.13.180083. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Naranje SM, Erali RA, Warner WC, Sawyer JR, Kelly DM. Epidemiology of Pediatric Fractures Presenting to Emergency Departments in the United States. J Pediatr Orthop. 2016;36(4):e45–8. doi: 10.1097/BPO.0000000000000595. [DOI] [PubMed] [Google Scholar]
- 12.DeWall M, Henderson CE, McKinley To, Phelps T, Dolan L, Marsh JL. Percutaneous reduction and fixation of displaced intra-articular calcaneus fractures. J Orthop Trauma. 2010;24(8):466–72. doi: 10.1097/BOT.0b013e3181defd74. [DOI] [PubMed] [Google Scholar]
- 13.Frick SL, Mehlman CT. The community orthopaedic surgeon taking trauma call: Pediatric supracondylar humeral fracture pearls and pitfalls. J Orthop Trauma. 2017;31(Suppl 6):S11–S15. doi: 10.1097/BOT.0000000000001012. [DOI] [PubMed] [Google Scholar]
- 14.Liu RW, Roocroft J, Bastrom T, Yaszay B. Surgeon learning curve for pediatric supracondylar humerus fractures. J Pediatr Orthop. 2011;31(8):818–24. doi: 10.1097/BPO.0b013e3182306884. [DOI] [PubMed] [Google Scholar]
- 15.Qudsi RA, Roberts HJ, Bhashyam AR, Losina E, Bae DS, Alexis F, Dyer GS. A self-reported needs assessment survey of pediatric orthopaedic education in Haiti. J Surg Educ. 2018;75(1):140–6. doi: 10.1016/j.jsurg.2017.06.007. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Shasti M, Li TP, Case AL, Hariharan AR, Jauregui JJ, Abzug JM. Cost analysis of treating pediatric supracondylar humerus fractures in community hospitals compared with a tertiary care hospital. J Am Acad Orthop Surg. 2019. Epub ahead of print. [DOI] [PubMed]
- 17.Taylor LK, Thomas GW, Karam MD, Kreiter CD, Anderson DD. Assessing wire navigation performance in the operating room. J Surg Educ. 2016;73(5):780–7. doi: 10.1016/j.jsurg.2016.03.018. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Taylor LK, Thomas GW, Karam MD, Kreiter CD, Anderson DD. Developing an objective assessment of surgical performance from operating room video and surgical imagery. IISE Trans Healthc Syst Eng. 2018;88(2):110–6. doi: 10.1080/24725579.2017.1418767. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Long S, Thomas GW, Anderson DD. Designing an affordable wire navigation surgical simulator. J Med Device. 2016;10(3):030921. doi: 10.1115/1.4033799. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Long S, Thomas GW, Anderson DD. An extensible orthopaedic wire navigation simulation platform. J Med Device. 2019;13(3):031001–310017. doi: 10.1115/1.4043461. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Tatum M, Thomas GW, Anderson DD. Design of a percutaneous articular fracture reduction simulator, in Des Med Devices Conf. [DOI] [PMC free article] [PubMed]
- 22.Long SA, Thomas G, Karam MD, Anderson DD. Do skills acquired from training with a wire navigation simulator transfer to a mock operating room environment? Clin Orthop Relat Res. 2019;477(10):2189–98. doi: 10.1097/CORR.0000000000000799. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Martin JA, Regehr G, Reznick R, MacRae H, Murnaghan J, Hutchison C, Brown M. Objective structured assessment of technical skill (OSATS) for surgical residents. Br J Surg. 1997;84(2):273–8. doi: 10.1046/j.1365-2168.1997.02502.x. [DOI] [PubMed] [Google Scholar]
- 24.Putnam MD, Adams JE, Lender P, Van Heest AE, Shanedling JR, Nuckley DJ, Bechtold JE. Examination of skill acquisition and grader bias in a distal radius fracture fixation model. J Surg Educ. 2018;75(5):1299–308. doi: 10.1016/j.jsurg.2018.01.019. [DOI] [PubMed] [Google Scholar]
- 25.Anderson DD, Long S, Thomas GW, Putnam MD, Bechtold JE, Karam MD. Objective Structured Assessments of Technical Skills (OSATS) does not assess the quality of the surgical result effectively. Clin Orthop Relat Res. 2016;474(4):874–81. doi: 10.1007/s11999-015-4603-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Karam MD, Thomas GW, Koehler DM, Westerlind BO, Lafferty PM, Ohrt GT, Marsh JL, Van Heest AE, Anderson DD. Surgical coaching from head-mounted video in the training of fluoroscopically guided articular fracture surgery. J Bone Joint Surg Am. 2015;97(12):1031–9. doi: 10.2106/JBJS.N.00748. [DOI] [PubMed] [Google Scholar]
- 27.Baumgaertner MR, Curtin SL, Lindskog DM, Keggi JM. The value of the tip-apex distance in predicting failure of fixation of peritrochanteric fractures of the hip. J Bone Joint Surg Am. 1995;77(7):1058–64. doi: 10.2106/00004623-199507000-00012. [DOI] [PubMed] [Google Scholar]
- 28.Baumgaertner MR, Solberg BD. Awareness of tip-apex distance reduces failure of fixation of trochanteric fractures of the hip. J Bone Joint Surg Br. 1997;79(6):969–71. doi: 10.1302/0301-620x.79b6.7949. [DOI] [PubMed] [Google Scholar]
- 29.Pennock AT, Charles M, Moor M, Bastrom TP, Newton PO. Potential causes of loss of reduction in supracondylar humerus fractures. J Pediatr Orthop. 2014;34(7):691–7. doi: 10.1097/BPO.0000000000000154. [DOI] [PubMed] [Google Scholar]
- 30.Chen C, White L, Kowalewski T, Aggarwal R, Lintott C, Comstock B, Kuksenok K, Aragon C, Holst D, Lendvay T. Crowd-Sourced Assessment of Technical Skills: A novel method to evaluate surgical performance. J Surg Res. 2014;187(1):65–71. doi: 10.1016/j.jss.2013.09.024. [DOI] [PubMed] [Google Scholar]
- 31.Dai JC, Lendvay TS, Sorensen MD. Crowdsourcing in surgical skills acquisition: A developing technology in surgical education. J Grad Med Educ. 2017;9(6):697–705. doi: 10.4300/JGME-D-17-00322.1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Kowalewski TM, Comstock B, Sweet R, Schaffhausen C, Menhadji A, Averch T, Box G, Brand T, Ferrandino M, Kaouk J, Knudsen B, Landman J, Lee B, Schwartz BF, McDougall E, Lendvay TS. Crowd-Sourced Assessment of Technical Skills for validation of basic laparoscopic urologic skills tasks. J Urol. 2016;195(6):1859–65. doi: 10.1016/j.juro.2016.01.005. [DOI] [PubMed] [Google Scholar]
- 33.Ghani KR, Miller DC, Linsell S, Brachulis A, Lane B, Sarle R, Dalela D, Menon M, Comstock B, Lendvay TS, Montie J, Peabody JO. Michigan Urological Surgery Improvement C. Measuring to Improve: Peer and crowd-sourced assessments of technical skill with robot-assisted radical prostatectomy. Eur Urol. 2016;69(4):547–50. doi: 10.1016/j.eururo.2015.11.028. [DOI] [PubMed] [Google Scholar]
- 34.Singer BD, Corbridge TC, Schroedl CJ, Wilcox JE, Cohen ER, McGaghie WC, Wayne DB. First-year residents outperform third-year residents after simulation-based education in critical care medicine. Simul Healthc. 2013;8(2):67–71. doi: 10.1097/SIH.0b013e31827744f2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Blum RH, Boulet JR, Cooper JB, Muret-Wagstaff SL. Harvard Assessment of Anesthesia Resident Performance Research G. Simulation-based assessment to identify critical gaps in safe anesthesia resident performance. Anesthesiology. 2014;120(1):129–41. doi: 10.1097/ALN.0000000000000055. [DOI] [PubMed] [Google Scholar]
- 36.Johnson Z, Thomas G, Long SA, Anderson DD. A pediatric supracondylar humerus fracture wire navigation simulator, in Des Med Devices Conf.
- 37.Tatum M. Simulating articular fracture reduction for surgical training, in Biomedical Engineering. Iowa City, IA: University of Iowa; 2019. p. 101. [Google Scholar]
- 38.Hattie J, Timperley H. The power of feedback. Review of Educational Research. 2007;77(1):81–112. [Google Scholar]
- 39.Hattie J. Influences on student learning. Inaugural lecture, University of Auckland, given on August 2, 1999. Available from: https://cdn.auckland.ac.nz/assets/education/about/research/documents/influences-on-student-learning.pdf.
- 40.Shute VJ. Focus on formative feedback. Review of Educational Research. 2008;78(1):153–89. [Google Scholar]
- 41.Taras M. The use of tutor feedback and student self-assessment in summative assessment tasks: Towards transparency for students and for tutors. Assessment & Evaluation in Higher Education. 2001;26(6):605–14. [Google Scholar]
- 42.Van der Kleij FM, Feskens RC, Eggen TJ. Effects of feedback in a computer-based learning environment on students’ learning outcomes: A meta-analysis. Review of educational research. 2015;85(4):475–511. [Google Scholar]
- 43.Narciss S. Springer.; Conditions and effects of feedback viewed through the lens of the interactive tutoring feedback model, in Scaling up assessment for learning in higher education. pp. 173–89. [Google Scholar]
- 44.Lee JD, See KA. Trust in automation: designing for appropriate reliance. Hum Factors. 2004;46(1):50–80. doi: 10.1518/hfes.46.1.50_30392. [DOI] [PubMed] [Google Scholar]
- 45.Kowalewski TM. University of Washington; Real-time Quantitative Assessment of Surgical Skill, in Ph.D. Thesis. Department of Electrical Engineering. [Google Scholar]
- 46.Kowalewski TM, White LW, Lendvay TS, Jiang IS, Sweet R, Wright A, Hannaford B, Sinanan MN. Beyond task time: automated measurement augments fundamentals of laparoscopic skills methodology. J Surg Res. 2014;192(2):329–38. doi: 10.1016/j.jss.2014.05.077. [DOI] [PubMed] [Google Scholar]
- 47.Peters JH, Fried GM, Swanstrom LL, Soper NJ, Sillin LF, Schirmer B, Hoffman K, Committee SF. Development and validation of a comprehensive program of education and assessment of the basic fundamentals of laparoscopic surgery. Surgery. 2004;135(1):21–7. doi: 10.1016/s0039-6060(03)00156-9. [DOI] [PubMed] [Google Scholar]
- 48.Lendvay TS, White L, Kowalewski T. Crowd-sourcing to assess surgical skill. JAMA Surg. 2015;150(11):1086–7. doi: 10.1001/jamasurg.2015.2405. [DOI] [PubMed] [Google Scholar]
- 49.Holst D, Kowalewski TM, White LW, Brand TC, Harper JD, Sorenson MD, Kirsch S, Lendvay TS. Crowd-Sourced Assessment of Technical Skills: An adjunct to urology resident surgical simulation training. J Endourol. 2015;29(5):604–9. doi: 10.1089/end.2014.0616. [DOI] [PubMed] [Google Scholar]
- 50.Holst D, Kowalewski TM, White LW, Brand TC, Harper JD, Sorensen MD, Truong M, Simpson K, Tanaka A, Smith R, Lendvay TS. Crowd-Sourced Assessment of Technical Skills: Differentiating animate surgical skill through the wisdom of crowds. J Endourol. 2015;29(10):1183–8. doi: 10.1089/end.2015.0104. [DOI] [PubMed] [Google Scholar]
- 51.Lendvay TS, Ghani KR, Peabody JO, Linsell S, Miller DC, Comstock B. PD46-03: Is crowdsourcing surgical skill assessment reliable? An analysis of robotic prostatectomies. J Urol. 2017;197(4S):E890–E1. [Google Scholar]
- 52.Kowalewski T, Sweet R, Menhadji A, Averch T, Box G, Brand T, Fearrandino M, Kaouk J, Knudsen B, Landman J, Lee B, Schwartz B, Comstock B, Schaffhausen C, McDougall E, Lendvay T. PD19-05: High-volume assessment of surgical videos via crowd-sourcing: The basic laparoscopic urologic skills (BLUS) Initiative. J Urol. 2015;193(4S):e393. [Google Scholar]
- 53.Celi LA, Ippolito A, Montgomery RA, Moses C, Stone DJ. Crowdsourcing knowledge discovery and innovations in medicine. J Med Internet Res. 2014;16(9):e216. doi: 10.2196/jmir.3761. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 54.Katz AJ. The role of crowdsourcing in assessing surgical skills. Surg Laparosc Endosc Percutan Tech. 2016;26(4):271–7. doi: 10.1097/SLE.0000000000000293. [DOI] [PubMed] [Google Scholar]
- 55.White LW, Kowalewski T, Brand T, Harper J, Comstock B, Hannaford B, Lendvay TS. PD6-08: Crowd-Sourced Assessment of Technical Skills (C-SATS™): Fast, economical and accurate assessment of robotic surgery. J Urol. 2014;191(4S) [Google Scholar]
- 56.White LW, Kowalewski TM, Dockter RL, Comstock B, Hannaford B, Lendvay TS. Crowd-Sourced Assessment of Technical Skill: A Valid Method for Discriminating Basic Robotic Surgery Skills. J Endourol. 2015;29(11):1295–301. doi: 10.1089/end.2015.0191. [DOI] [PubMed] [Google Scholar]
- 57.Dirschl DR, Marsh JL, Buckwalter JA, Gelberman R, Olson SA, Brown TD, Llinias A. Articular fractures. J Am Acad Orthop Surg. 2004;12(6):416–23. doi: 10.5435/00124635-200411000-00006. [DOI] [PubMed] [Google Scholar]
