Abstract
Background
Our aim was to evaluate the impact of a computerized echocardiographic simulator on the learning curve for transesophageal echocardiography (TEE) hemodynamic assessment of ventilated patients in the ICU.
Methods
We performed a prospective study in two university hospital medical ICUs. Using our previously validated skill assessment scoring system (/40 points), we compared learning curves obtained with (interventional group, n = 25 trainees) and without (control group, n = 31 trainees) use of a simulator in the training. Three evaluations were performed after 1 (M1), 3 (M3) and 6 months (M6) while performing two TEE examinations graded by an expert. Competency was defined as a score >35/40.
Results
Competency was achieved after an average of 32.5 ± 10 supervised studies in the control group compared with only 13.6 ± 8.5 in the interventional group (p < 0.0001). At M6, a significant between-group difference in number of supervised TEE was observed (17 [14–28] in the control group vs. 30.5 [21.5–39.5] in the interventional group, p = 0.001). The score was significantly higher in the interventional group at M1 (32.5 [29.25–35.5] vs. 24.75 [20–30.25]; p = 0.0001), M3 (37 [33.5–38.5] vs. 32 [30.37–34.5]; p = 0.0004), but not at M6 (37.5 [33–39] vs. 36 [33.5–37.5] p = 0.24).
Conclusion
Inclusion of echocardiographic simulator sessions in a standardized curriculum may improve the learning curve for hemodynamic evaluation of ventilated ICU patients.
Electronic supplementary material
The online version of this article (doi:10.1186/s13613-016-0132-x) contains supplementary material, which is available to authorized users.
Keywords: Education, Transesophageal echocardiography, Critical care, Echocardiography, Mannequin echocardiographic simulator
Background
Since the early 1960s and the first resuscitation manikin, technological improvements have allowed the development of virtual-reality training simulators [1, 2]. In the surgical field, the impact of 3D haptic laparoscopy simulators has been extensively explored, and the transferability of bedside skills after simulation training [3–5] has been validated, leading to a European consensus on a competency-based virtual-reality training program [6].
Critical care echocardiography (CCE) has gained acceptance and in its basic form based on transthoracic echocardiography (TTE) is now recommended for inclusion in the curriculum of all intensivists [7, 8]. Advanced CCE is also recognized as a way to monitor hemodynamics fully in the ICU [9], but requires acquisition of technical skills in transesophageal echocardiography (TEE). Traditionally, after a standard curriculum using didactic courses, trainees acquire technical skills at the bedside, under the supervision of experienced physicians. International consensus statements establish how physicians can acquire the competency needed for ICU practice [8, 9]. Types of echocardiographic views, measurements and the overall number of bedside TEE and TTE examinations needed to achieve competency are defined, but very little is mentioned about virtual-reality simulation [10]. To our knowledge, there is only one study reporting its inclusion in a cardiovascular curriculum, in terms of skills acquisition and proficiency [11].
In two previous studies we validated a skills assessment scoring system and used it to demonstrate that at least 31 supervised bedside TEE examinations, performed during a 6-month period, were required to achieve competency in hemodynamic evaluation of ventilated ICU patients [12, 13]. The third part of this educational process is represented herein and is designed to evaluate the impact of a TEE virtual-reality simulator, integrated into the previously described curriculum.
Methods
We conducted a prospective, multicenter study in two French university hospital medical ICUs between May 2012 and November 2014. The results for trainees trained without a simulator (control group) between November 2006 and June 2010 [13] were compared with those recorded during the second period between May 2012 and November 2014 with the simulator (interventional group). In the participating centers, TEE has been used for years as a first-line tool for hemodynamic assessment in critically ill patients. As a result, more than 350–450 TEE examinations are performed yearly in each center in mechanically ventilated patients presenting with shock or acute respiratory failure. Our study was therefore considered as part of routine practice by our local ethics committee, and no informed consent was required from the patients or their next of kin.
All volunteer residents rotating in our ICUs each 6 months, without previous experience in TEE, were consecutively included and constituted the evaluated population. Their previous experience in echocardiography was graded as level 0 (no experience at all) or level 1 (previous experience in TTE without TEE experience). Each trainee performed and interpreted TEE examinations online, under the supervision of an expert, and the number of supervised TEE examinations was prospectively recorded during a 6-month period, as previously done [12]. A 2-h didactic course on echocardiographic basic ultrasounds was performed by an expert for the two groups (control group and interventional group) at the beginning of each 6-month period. Trainees in the interventional group also had two 3-h individual sessions of practical hands-on training using an echocardiographic simulator (Vimedix CAE Healthcare Inc, Montréal, Canada or Heartworks Intensive Medical Ltd, London, UK) during the first 3 months. Simulator learning was focused on the acquisition of the main esophageal views used for hemodynamic evaluation at the bedside: the mid-esophagus long-axis view 0°–120°, the transgastric short-axis view 0°–110° and the upper-esophagus (great vessels) view 0°–90° (Additional file 1: Figure S1).
After 1 (M1), 3 (M3) and 6 months (M6), trainees were evaluated by a supervisor using our previously validated scoring system (Table 1) [12, 13]. Briefly, the maximum score is 40 points with four fields of skills to grade: practical skill (/14 points) reflecting the ability of the trainee to obtain standard TEE views, evaluation skill (/10 points) assessing the semiquantitative evaluation of right ventricular size, respiratory variations in the diameter of the superior vena cava, mitral and aortic regurgitation, and pericardial effusion, technical skill (/8 points) reflecting the ability of the trainee accurately to measure simple hemodynamic parameters (e.g., left ventricular ejection fraction, velocity–time integral of left ventricular outflow tract Doppler velocities), and finally the interpretation skill (/8 points), which referred to the trainee’s ability to summarize the information obtained by the TEE examination and to suggest adequate therapeutic changes accordingly.
Table 1.
Score | ||||
---|---|---|---|---|
Practical skill | ||||
Introduction of the probe | No | Problematic | Yes | /2 |
TE long-axis view at 0° | Not recorded | Not optimal | Optimal | /2 |
TE long-axis view at 120° | Not recorded | Not optimal | Optimal | /2 |
TG short-axis view at 0° | Not recorded | Not optimal | Optimal | /2 |
TG short-axis view at 120° | Not recorded | Not optimal | Optimal | /2 |
TE view of the base of the heart at 0° | Not recorded | Not optimal | Optimal | /2 |
TE view of the base of the heart at 90° | Not recorded | Not optimal | Optimal | /2 |
Total | /14 | |||
Evaluation skill | ||||
Mitral regurgitation | None | Moderate | Marked to massive | /2 |
Aortic regurgitation | None | Moderate | Marked to massive | /2 |
Dilatation of right ventricle | None | Moderate | Marked | /2 |
Pericardial effusion | None | Non compressive | Compressive | /2 |
Variations in diameter of superior vena cava | None | Minimal | Large | /2 |
Total | /10 | |||
Intensivist | Expert | |||
Technical skill | ||||
E/A ratio | /2 | |||
LV FAC (%) | /2 | |||
Aortic VTI (cm) | /2 | |||
Pulmonary VTI (cm) | /2 | |||
Total | /8 | |||
Interpretation skill | ||||
LV contractility | Normal | Moderately decreased | Greatly decreased | /2 |
Hypovolemia | No | Yes | /2 | |
RV failure | No | Yes | /2 | |
Treatment proposed | Wrong or incomplete | Right | /2 | |
Total | /8 | |||
Final score/40 |
TE transesophageal, TG transgastric, FAC fractional area change, LV left ventricle, RV right ventricle, VTI velocity–time integral
The supervisor was an expert in advanced CCE and a full-time intensivist with more than 10 years of ICU experience and >200 TEE/year examinations performed on mechanically ventilated patients. All evaluations were performed by the same expert in each center, and trainees did not receive any feedback or assistance while performing TEE until they wrote the echocardiographic report.
Statistical analysis
Statistical analysis was performed using MedCalc (9030 Mariakerke, Belgium). Continuous variables were expressed as median [25–75 percentiles]. Between-group comparisons were performed using nonparametric tests. ANOVA was used to compare changes in score and to build the learning curve. We compared learning curves between the previous period with no simulator use in the training process, in which we included 31 residents [13], i.e., the control group, and the current period using the simulator, i.e., the interventional group. In the control group, we excluded from the analysis 10 trainees from our previous published study performed in 41 trainees [13] since in one center of this previous study no TEE simulator was available for the second period. We were then able to compare the performance of trainees in the two groups with the same process of evaluation, the same experts, as well as the same scoring system. In the two groups, competency was arbitrarily defined by a skills assessment score >35/40, as previously done [13].
Results
Fifty-six trainees were included during the two periods, 31 in the control group and 25 in the interventional group. Trainee characteristics were similar in the two groups. All were residents in the interventional group, while 30 were residents and 1 was senior intensivist in the control group (p = 0.58). No difference was observed in trainees’ specialty of origin (p = 0.39). Most were anesthesiologists (62 %) and cardiologists (19 %) in the interventional group. Eight trainees (32 %) had echocardiography level 1 experience in the interventional group and 15 (48 %) in the control group (p = 0.38).
The mean score improved faster over the 6-month period in the interventional group (p = 0.046, Fig. 1), with a significantly higher score at M1 and M3 although not at M6 (Table 2). Despite a significantly lower number of supervised TEE examinations at M6 (17 [14–28] versus 30.5 [21.5–39.5], p = 0.0004, Table 2), trainees in the interventional group reached the same average score as in the control group (37.5 [33–39] versus 36 [33.5–37.5], respectively, p = 0.24, Table 2). Competency was obtained after an average of 32.5 ± 10 supervised studies in the control group compared with 13.6 ± 8.5 in the interventional group (p < 0.0001). Comparison between each part of the scoring system is depicted in Table 3 and in Fig. 2. Practical and technical skills were significantly increased in the interventional group at M1 and M3, although not at M6, whereas no difference was observed for the evaluation and interpretation skills, except at M1.
Table 2.
M1 | M3 | M6 | ANOVA | |
---|---|---|---|---|
Mean score, interventional group | 32.5 [29.25–35.5] | 37 [33.5–38.5] | 37.5 [33–39] | p = 0.048 |
Mean score, control group | 24.75 [20–30.25] | 32 [30.37–34.5] | 36 [33.5–37.5] | |
p | <0.0001 | 0.0004 | 0.24 | |
Number of supervised TEE examinations, interventional group | 4.5 [3.5–6.25] | 10.5 [8.5–13.25] | 17 [14–28] | p = 0.001 |
Number of supervised TEE examinations, control group | 5.5 [3.75–6.5] | 15.5 [14–20] | 30.5 [21.5–39.5] | |
p | 0.31 | 0.0003 | 0.0004 |
Skills assessment score with the number of supervised TEE examinations at 1 (M1), 3 (M3) and 6 months (M6) in the interventional (with simulator) and control (without simulator) groups. ANOVA column corresponds to analysis of changes from M1 to M6, whereas p line relates to comparison between both groups at each evaluation. Data are expressed as median [25–75 percentiles]
TEE transesophageal echocardiography
Table 3.
M1 | M3 | M6 | ||||
---|---|---|---|---|---|---|
Interventional | Control | Interventional | Control | Interventional | Control | |
Practical skills (/14 pts) | 11* [10–11.5] | 8.5 [7–10.5] | 12* [11.5–13]* | 11.5 [10.5–12] | 13 [12.5–13.5] | 13 [12–14] |
Evaluation skills (/10 pts) | 8* [7.5 – 9.5] | 7.5 [6.5–8.5] | 9.5 [9, 10] | 9 [8.5–9.5] | 9.5 [9–9.5] | 9 [9, 10] |
Technical skills (/8 pts) | 5* [3.5 – 6] | 3 [2–4] | 6* [5.5–7] | 5 [4–5.5] | 6.5 [5–7] | 6 [5.5–7] |
Interpretation skills (/8 pts) | 7* [5, 6, 6–8] | 5 [4–7.5] | 8 [7, 8] | 8 [7, 8] | 8 [8] | 8 [6–8] |
Interventional group regards the group with a simulator and the control group the period without a simulator. Data are expressed as median [25–75 percentiles]
* p < 0.05 versus control group
At M1, M3 and M6 there was no between-group difference in the procedural time duration (interventional vs. control: 20 [17.5–24.5] vs. 19.5 min [17–25], 15 [12.5–18] vs. 14.75 min [13–18] and 12.75 [11.5–15.5] vs. 12 min [10–16], respectively).
Discussion
Our study demonstrates that the addition of virtual-reality simulator sessions to a standardized 6-month curriculum improves the learning curve for ICU TEE hemodynamic assessment. The use of the simulator especially reduced the number of TEE examinations required to achieve competency by improving acquisition of practical and technical skills.
Virtual-reality simulators have proven to be an efficient addition to curricula in various healthcare specialties, including laparoscopy [14, 15], upper- and lower-gut endoscopy [16, 17] and endoscopic sinus surgery [18]. Most recent virtual-reality echocardiographic simulators are quite realistic, according to hands-on training or questionnaire evaluations [19, 20]. In a questionnaire study, a vast majority of participants answered that simulators are realistic, easy-to-use and helpful for image acquisition and interpretation [21].
TEE is a complex endeavor, involving motor and cognitive skills. In a study of 18 senior anesthesiologists without any ultrasound knowledge, Matyal et al. examined the impact of web-based ultrasound didactics and biweekly 90-min hands-on sessions with a TEE simulator for 4 weeks (13 target cut planes were taught). Weekly evaluation of practical skills with kinematic analysis of probe motion [22] depicted a progressive decrease in peak movements and path lengths over the 4 weeks of training, and the results at the final evaluation were close to those obtained by the experts. This three-dimensional anatomic approach explains the higher score observed in our interventional group, especially supported by an increase in practical skills. The possibility of obtaining a double image simultaneously showing the ultrasound image and the anatomic representation with the ultrasound beam also enables trainees to assimilate more accurately probe movement skills and probe positioning regarding the surrounding anatomic structures [23, 24].
A few studies of the bedside transferability of acquired skills have suggested that a TEE simulator could improve the training process [19, 20, 25–27], but were unclear about the real ability of physicians to obtain and correctly interpret images over a large range of pathological cases in real patients. They were mainly done as pretest/posttest procedures, probably due to the lack of any scoring system for evaluation of trainees at the bedside. Even a randomized study in 46 anesthesia residents (80-min TTE training using a simulator after videos and tutorial) found no significant improvement in performance in image acquisition and posttest evaluation when using a simulator [20]. Only one randomized study performed in 42 anesthesia residents reported significantly higher image quality in ten preselected standard views in real patients, especially for younger residents, after using the simulator [28]. Thanks to our previously validated scoring system, which was reported as discriminatory and sensitive to change [12], our study shows that a TEE simulator could be used to train intensivists in the use of TEE in the ICU since all evaluations were performed on real patients with hemodynamic instability. Interestingly, although we mainly reported a significant improvement in the interventional group for acquisition of practical skills, i.e., the ability of trainees to record the main views, we also found a significant improvement in technical skills, i.e., the ability of trainees to measure simple hemodynamic parameters accurately, which makes sense since technical skills are very frequently related to practical skills. These two aspects represent little more than half of the scoring system (22/40) and, interestingly, we found no significant difference between the interventional and control groups in the evaluation or interpretation skills, which are related to the ability of trainees to interpret the examination adequately and propose the right treatment. We may assume that these two components are less reliable with the use of a simulator, especially with our approach in which we used the simulator to help in the acquisition of normal images in the absence of pathology.
How inclusion of simulation will impact on a standardized curriculum has yet to be evaluated formally. The huge between-group difference we observed in the number of supervised TEE examinations required for acquisition of competency was surprising. It can in part be explained by the fact that the practical and technical skills parts represent more than half of our overall scoring system, as discussed above. However, we cannot at this point recommend the use of only 14 supervised TEE examinations for acquisition of ICU TEE competency when using a simulator, when the current recommendation is 35 [13]. Our results only support the recommendation, made by international experts in the very last document of the European Society of Intensive Care Medicine, to include TEE simulation in the training process for advanced CCE [8]. Damp et al. [11] compared fellow cardiologists, some of whom completed standard TEE training and others who were also trained with TEE simulator, while they performed bedside TEE on real patients. Like us, they found that cardiologists in the simulator group acquired a very significantly higher number of views without any assistance with much lower variability among trainees [11]. Damp et al. [11] also observed a trend toward a shorter overall TEE duration for cardiologists trained with the simulator, whereas there was no correlation between the time spent on the simulator and the evaluation scores. Previous studies on TEE simulation considered that a median 1-h simulator training is sufficient to note significant results in skills acquisition [20, 25, 29, 30]. The study by Damp et al., like ours, should help define in the future how simulation sessions have to be included in the training program of diplomas in advanced CCE.
Our study has some limitations. First, the interventional and control groups were not studied during the same period, since we compared the current period using a simulator with a previous evaluation of training not using a simulator [13]. We can’t exclude that trainees in the interventional group would have greater baseline knowledge of ultrasound than trainees in the control group included several years before. However, we used exactly the same validated scoring system to grade the ability of trainees to use and interpret ICU TEE examinations for hemodynamic evaluation. Moreover, the same expert centers also participated in the study with the same supervisors in the two periods. Second, as a consequence of the first limitation, we did not randomize trainees in the two groups. However, as shown in the results, trainees did not differ between the two periods, but we cannot exclude recruiting bias of trainees and also of the patients evaluated to grade the trainees. Third, although the simulator improved training, it acted mainly as an “accelerator” since at the end of the 6 months of training we did not note any difference between the two groups. Fourth, the supervisors have participated in training the residents and there could be a potential bias in favor of a higher score in the interventional group. Finally, our scoring system was especially developed and validated for full hemodynamic evaluation at the bedside in critically ill patients and our results cannot be extrapolated to other more specific cardiological situations.
Conclusion
To the best of our knowledge, this is the first time that the inclusion of a TEE virtual-reality simulator in standardized ICU TEE training improved the learning curve of novices. Such training also reduced the number of supervised bedside TEE examinations needed to achieve competency in hemodynamic evaluation in the ICU. How and to what extent this kind of approach can be included in standardized curricula remain to be evaluated.
Authors’ contributions
GP and CC contributed to conception and design of the research; GP, CC and XR contributed to the acquisition of the data; all authors contributed to the analysis and interpretation of the data and drafting of the manuscript, critically revised the manuscript and agree to be fully accountable for ensuring the integrity and accuracy of the work. All authors read and approved the final manuscript.
Competing interests
The authors declare that they have no competing interests.
Additional file
Contributor Information
Gwénaël Prat, Email: gwenael.prat@chu-brest.fr.
Cyril Charron, Email: cyril.charron@aphp.fr.
Xavier Repesse, Email: xavier.repesse@aphp.fr.
Pierre Coriat, Email: pierre.coriat@psl.aphp.fr.
Pierre Bailly, Email: pierre.bailly@chu-brest.fr.
Erwan L’her, Email: erwan.lher@chu-brest.fr.
Antoine Vieillard-Baron, Email: antoine.vieillard-baron@aphp.fr.
References
- 1.Rosen KR. The history of medical simulation. J Crit Care. 2008;23(2):157–166. doi: 10.1016/j.jcrc.2007.12.004. [DOI] [PubMed] [Google Scholar]
- 2.Good ML. Patient simulation for training basic and advanced clinical skills. Med Educ. 2003;37:14–21. doi: 10.1046/j.1365-2923.37.s1.6.x. [DOI] [PubMed] [Google Scholar]
- 3.Stefanidis D, Scerbo MW, Montero PN, Acker CE, Smith WD. Simulator training to automaticity leads to improved skill transfer compared with traditional proficiency-based training: a randomized controlled trial. Ann Surg. 2012;255(1):30–37. doi: 10.1097/SLA.0b013e318220ef31. [DOI] [PubMed] [Google Scholar]
- 4.Sturm LP, Windsor JA, Cosman PH, Cregan P, Hewett PJ, Maddern GJ. A systematic review of skills transfer after surgical simulation training. Ann Surg. 2008;248(2):166–179. doi: 10.1097/SLA.0b013e318176bf24. [DOI] [PubMed] [Google Scholar]
- 5.Gurusamy K, Aggarwal R, Palanivelu L, Davidson BR. Systematic review of randomized controlled trials on the effectiveness of virtual reality training for laparoscopic surgery. Br J Surg. 2008;95(9):1088–1097. doi: 10.1002/bjs.6344. [DOI] [PubMed] [Google Scholar]
- 6.Van Dongen KW, Ahlberg G, Bonavina L, Hyltander A, Schijven MP, Stefani A, Van Der Zee DC, Broeders IA. European consensus on a competency-based virtual reality training program for basic endoscopic surgical psychomotor skills. Surg Endosc. 2011;25(1):166–171. doi: 10.1007/s00464-010-1151-6. [DOI] [PubMed] [Google Scholar]
- 7.Mayo PH, Beaulieu Y, Doelken P, Feller-Kopman D, Harrod C, Kaplan A, Oropello J, Vieillard-Baron A, Axler O, Lichtenstein D, Maury E, Slama M, Vignon P. American College of Chest Physicians/La Societe de Reanimation de Langue Francaise statement on competence in critical care ultrasonography. Chest. 2009;135(4):1050–1060. doi: 10.1378/chest.08-2305. [DOI] [PubMed] [Google Scholar]
- 8.International expert statement on training standards for critical care ultrasonography. Intensive Care Med. 2011;37:1077–83. [DOI] [PubMed]
- 9.International consensus statement on training standards for advanced critical care echocardiography. Intensive Care Med. 2014;40(5):654–666. [DOI] [PubMed]
- 10.Clau-Terré F, Vivek S, Cholley B, Gonzales-Ajulas T, Galinanes M, Evangelista A, Fletcher N. Can simulation help to answer the demand for echocardiography education? Anesthesiology. 2014;120:32–41. doi: 10.1097/ALN.0000000000000072. [DOI] [PubMed] [Google Scholar]
- 11.Damp J, Anthony R, Davidson MA, Mendes L. Effects of transesophageal echocardiography simulator training on learning and performance in cardiovascular medicine fellows. J Am Soc Echocardiogr. 2013;26(12):1450–1456. doi: 10.1016/j.echo.2013.08.008. [DOI] [PubMed] [Google Scholar]
- 12.Charron C, Prat G, Caille V, Belliard G, Lefevre M, Aegerter P, Boles JM, Jardin F, Vieillard-Baron A. Validation of a skills assessment scoring system for transesophageal echocardiographic monitoring of hemodynamics. Intensive Care Med. 2007;33(10):1712–1718. doi: 10.1007/s00134-007-0801-1. [DOI] [PubMed] [Google Scholar]
- 13.Charron C, Vignon P, Prat G, Tonnelier A, Aegerter P, Boles JM, Amiel JB, Vieillard-Baron A. Number of supervised studies required to reach competence in advanced critical care transesophageal echocardiography. Intensive Care Med. 2013;39(6):1019–1024. doi: 10.1007/s00134-013-2838-7. [DOI] [PubMed] [Google Scholar]
- 14.Larsen CR, Soerensen JL, Grantcharov TP, Dalsgaard T, Schouenborg L, Ottosen C, Schroeder TV, Ottesen BS. Effect of virtual reality training on laparoscopic surgery: randomised controlled trial. BMJ. 2009;338:b1802. doi: 10.1136/bmj.b1802. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Rosser JC, Jr, Colsant BJ, Lynch PJ, Herman B, Klonsky J, Young SM. The use of a “hybrid” trainer in an established laparoscopic skills program. JSLS. 2006;10(1):4–10. [PMC free article] [PubMed] [Google Scholar]
- 16.Sedlack RE. Validation of computer simulation training for esophagogastroduodenoscopy: pilot study. J Gastroenterol Hepatol. 2007;22(8):1214–1219. doi: 10.1111/j.1440-1746.2007.04841.x. [DOI] [PubMed] [Google Scholar]
- 17.Sedlack RE, Baron TH, Downing SM, Schwartz AJ. Validation of a colonoscopy simulation model for skills assessment. Am J Gastroenterol. 2007;102(1):64–74. doi: 10.1111/j.1572-0241.2006.00942.x. [DOI] [PubMed] [Google Scholar]
- 18.Fried MP, Sadoughi B, Weghorst SJ, Zeltsan M, Cuellar H, Uribe JI, Sasaki CT, Ross DA, Jacobs JB, Lebowitz RA, Satava RM. Construct validity of the endoscopic sinus surgery simulator: II. Assessment of discriminant validity and expert benchmarking. Arch Otolaryngol Head Neck Surg. 2007;133(4):350–357. doi: 10.1001/archotol.133.4.350. [DOI] [PubMed] [Google Scholar]
- 19.Bick JS, Demaria S, Jr, Kennedy JD, Schwartz AD, Weiner MM, Levine AI, Shi Y, Schildcrout JS, Wagner CE. Comparison of expert and novice performance of a simulated transesophageal echocardiography examination. Simul Healthc. 2013;8(5):329–334. doi: 10.1097/SIH.0b013e31829068df. [DOI] [PubMed] [Google Scholar]
- 20.Edrich T, Seethala RR, Olenchock BA, Mizuguchi AK, Rivero JM, Beutler SS, Fox JA, Liu X, Frendl G. Providing initial transthoracic echocardiography training for anesthesiologists: simulator training is not inferior to live training. J Cardiothorac Vasc Anesth. 2014;28(1):49–53. doi: 10.1053/j.jvca.2013.07.011. [DOI] [PubMed] [Google Scholar]
- 21.Platts DG, Humphries J, Burstow DJ, Anderson B, Forshaw T, Scalia GM. The use of computerised simulators for training of transthoracic and transoesophageal echocardiography. The future of echocardiographic training? Heart Lung Circ. 2012;21(5):267–274. doi: 10.1016/j.hlc.2012.03.012. [DOI] [PubMed] [Google Scholar]
- 22.Matyal R, Mitchell JD, Hess PE, Chaudary B, Bose R, Jainandunsing JS, Wong V, Mahmood F. Simulator-based transesophageal echocardiographic training with motion analysis: a curriculum-based approach. Anesthesiology. 2014;121(2):389–399. doi: 10.1097/ALN.0000000000000234. [DOI] [PubMed] [Google Scholar]
- 23.Matyal R, Bose R, Warraich H, Shahul S, Ratcliff S, Panzica P, Mahmood F. Emerging technology review. J Cardiothorac Vasc Anesth. 2011;25:177–181. doi: 10.1053/j.jvca.2010.08.008. [DOI] [PubMed] [Google Scholar]
- 24.Bose R, Matyal R, Panzica P, Karthik S, Subramaniam B, Pawlowski J, Mitchell J, Mahmood F. Transesophageal echocardiography simulator: a new learning tool. J Cardiothorac Vasc Anesth. 2009;23(4):544–548. doi: 10.1053/j.jvca.2009.01.014. [DOI] [PubMed] [Google Scholar]
- 25.Vegas A, Meineri M, Jerath A, Corrin M, Silversides C, Tait G. Impact of online transesophageal echocardiographic simulation on learning to navigate the 20 standard views. J Cardiothorac Vasc Anesth. 2013;27(3):531–535. doi: 10.1053/j.jvca.2012.09.006. [DOI] [PubMed] [Google Scholar]
- 26.Ogilvie E, Vlachou A, Edsell M, Fletcher SN, Valencia O, Meineri M, Sharma V. Simulation-based teaching versus point-of-care teaching for identification of basic transoesophageal echocardiography views: a prospective randomised study. Anaesthesia. 2015;70(3):330–335. doi: 10.1111/anae.12903. [DOI] [PubMed] [Google Scholar]
- 27.Neelankavil J, Howard-Quijano K, Hsieh TC, Ramsingh D, Scovotti JC, Chua JH, Ho JK, Mahajan A. Transthoracic echocardiography simulation is an efficient method to train anesthesiologists in basic transthoracic echocardiography skills. Anesth Analg. 2012;115(5):1042–1051. doi: 10.1213/ANE.0b013e318265408f. [DOI] [PubMed] [Google Scholar]
- 28.Ferrero NA, Bortsov AV, Arora H, Martinelli SM, Kolarczyk LM, Teeter EC, Zvara DA, Kumar PA. Simulator training enhances resident performance in transesophageal echocardiography. Anesthesiology. 2014;120(1):149–159. doi: 10.1097/ALN.0000000000000063. [DOI] [PubMed] [Google Scholar]
- 29.Jelacic S, Bowdle A, Togashi K, VonHomeyer P. The use of TEE simulation in teaching basic echocardiography skills to senior anesthesiology residents. J Cardiothorac Vasc Anesth. 2013;27(4):670–675. doi: 10.1053/j.jvca.2013.01.016. [DOI] [PubMed] [Google Scholar]
- 30.Bose RR, Matyal R, Warraich HJ, Summers J, Subramaniam B, Mitchell J, Panzica PJ, Shahul S, Mahmood F. Utility of a transesophageal echocardiographic simulator as a teaching tool. J Cardiothorac Vasc Anesth. 2011;25(2):212–215. doi: 10.1053/j.jvca.2010.08.014. [DOI] [PubMed] [Google Scholar]