Skip to main content
JSLS : Journal of the Society of Laparoscopic & Robotic Surgeons logoLink to JSLS : Journal of the Society of Laparoscopic & Robotic Surgeons
. 2018 Jan-Mar;22(1):e2017.00081. doi: 10.4293/JSLS.2017.00081

Head-to-Head Comparison of Three Virtual-Reality Robotic Surgery Simulators

Alexandria M Hertz 1,, Evalyn I George 2, Christine M Vaccaro 3, Timothy C Brand 4
PMCID: PMC5863693  PMID: 29618918

Abstract

Background and Objectives:

There are several different commercially available virtual-reality robotic simulators, but very little comparative data. We compared the face and content validity of 3 robotic surgery simulators and their pricing and availability.

Methods:

Fifteen participants completed one task on each of the following: dV-Trainer (dVT; Mimic Technologies, Inc., Seattle, Washington, USA), da Vinci Skills Simulator (dVSS; Intuitive Surgical Inc., Sunnyvale, California, USA), and RobotiX Mentor (RM; 3D Systems, Rock Hill, South Carolina, USA). Participants completed previously validated face and content validity questionnaires and a demographics questionnaire. Statistical analysis was then performed on the scores.

Results:

Participants had a mean age of 29.6 (range, 25–41) years. Most were surgical trainees, having performed a mean of 8.6 robotic primary surgeries. For face validity, ANOVA showed a significant difference favoring the dVSS over the dVT (P = .001), and no significant difference between the RM, dVSS, and dVT. Content validity revealed similar results, with a significant difference between the dVSS and dVT (P = .021), a trend toward a difference between the RM and dVT (P = .092), and no difference between the dVSS and RM (P = .99).

Conclusion:

All simulators demonstrated evidence of face and content validity, with significantly higher scores for the dVSS; it is also the least costly ($80,000 for the simulator), although it is frequently unavailable because of intra-operative use. The dVT and RM have similar face and content validity, are slightly more expensive, and are readily available.

Keywords: da Vinci robot, Robotics, Simulation, Virtual reality

INTRODUCTION

As robotic surgery becomes a more common choice for operative procedures because of faster recovery times, lower complication rates, decreased blood loss, and recovery of site-specific functions, surgeons are tasked with improving and retaining skill without affecting patient safety.16 Although medicine is moving away from intraoperative training, the high cost of the da Vinci Surgical System (Intuitive Surgical, Inc., Sunnyvale, CA), the most common commercially available robotic platform, impinges on their capacity for use outside of the operating room (OR). Considering the capacity for harm to the patient and several widely publicized malpractice suits involving the da Vinci, it is crucial that surgeons be adequately trained and maintain their skills.7

Virtual-reality (VR) simulators offer opportunities for surgical training outside of the OR as well as standardized skill evaluations.810 As institutions begin instituting their own training and credentialing curriculums for robotic surgery or look to adopting multiplatform curricula such as the Fundamentals of Robotic Surgery program1013, we present a recently conducted head-to-head comparison of 3 VR robotic simulators with regard to face and content validity, as well as a differential cost analysis.

There are currently 5 VR simulators that are commercially available for robotic training.14,15 This project focused on 3 that had already been validated. The dV Trainer (dVT) from Mimic Technologies, Inc. (Seattle, Washington, USA) and da Vinci Skills Simulator (dVSS) from Intuitive Surgical, Inc. (Sunnyvale, CA. USA) have both been the subject of many prior validation studies.1625 The RobotiX Mentor (RM) (3D Systems, Littleton, Colorado, USA) is the most recent commercially available VR simulator for robotic skills development and has shown evidence of face, content, and construct validity.26 The other 2 commercially available simulators are the RoSS Robotic Surgery Simulator (Simulated Surgical Systems LLC, San Jose, California, USA) and the SimSurgery Educational Platform (SEP) (SimSurgery, Oslo, Norway). The SEP is predominantly distributed in Europe; in addition, there has been limited validation of both of these platforms.27 The unavailability of the latter 2 in the United States restricts their use.

MATERIALS AND METHODS

Fifteen participants, ranging from senior medical students to experienced robotic systems surgeons, were recruited to participate in this observational, comparative study. Institutional Review Board deemed this project exempt, as was no intervention was involved and no identifiable personal information was gathered. The participants were asked to complete a general demographics questionnaire. They then completed 1 task on each of the 3 simulators. The task was of the participant's choosing and did not have to be consistent among the simulators. We chose this design to obtain a true gauge of the participants' opinions without the influence of mandated procedures. No time limits were set for the tasks and time spent on any particular simulator.

Following each of the tasks, the participants were asked to complete formerly validated questionnaires for evaluation of face and content validity (Figure 1).28 Face validity is considered to be the degree of realism between the system and the demonstrated activity. Alternatively, content validity illustrates the efficacy of a system as an instructional apparatus.

Figure 1.

Figure 1.

Face (top) and content (bottom) validation questionnaires.

Statistical analysis was performed by analysis of variance (ANOVA; SPSS version 19 software; IBM Analytics, Armonk, New York) of the scores obtained from the face and content questionnaires, with P < .05 indicating a significant difference. Descriptive analysis and means were performed for the demographics questionnaire when appropriate.

Commercial pricing for the different simulators was obtained by contacting representatives of each company via e-mail.

RESULTS

The 15 participants had an average age of 29.6 (range 25–41) years, and most were surgical residents with an average of 8.6 robotic primary cases (Table 1). All completed tasks on 3 robotic surgical simulators and completed a series questionnaires designed to illustrate face and content validity.

Table 1.

Characteristics of Surgical Patients

Characteristic Data
Age (y)
    Mean 29.6
    Range 24–41
Gender (n)
    Female 5
    Male 10
Years of training (n) 3
Medical student level
    Intern 4
    PGY2 4
    Senior resident 2
    Attending 2
Specialty
    General surgery 8
    Urology 7
    Gynecology 0
Average number of robotic cases (as primary surgeon)
    Mean 7.53
    Range 0–50
Average number of laparoscopic cases (as primary surgeon)
    Mean 17.5
    Range 0–100
Handedness
    Left 2/15
    Right 13/15
Musical experience
    Yes (n/total) 8/15
    Mean years played 12.6
Military deployment (n/total) 0/15
Mean days since last robotic case 168.5
No prior robotic experience (n/total) 9/15
Mean days since last simulation 8.2
No prior robotic simulation experience (n/total) 5/15

The average scores for the individual simulators' respective face and content validity questionnaires are as follows: dVSS, 27.2 /27.73; dVT, 21.4/23.33; and RM, 24.73/26.80. These means are based on total combined scores (range, 5–30), with higher scores equating to greater demonstrated validity (Table 2).

Table 2.

Questionnaire Scores

Robotic System First Questionnaire Mean Second Questionnaire Mean SE First Questionnaire SE Second Questionnaire
dVSS 27.2 27.733 0.812 0.771
dVT 21.4 23.33 1.341 1.53
RM 24.733 26.80 0.933 0.812

The face validation questionnaire showed the dVSS to be significantly more realistic when compared with dVT (P = .001). There was no difference between the dVSS and RM (P = .316). The RM showed a trend toward being more realistic in comparison to the dVT, although the difference was not significant (P = .092) (Table 3).

Table 3.

Questionnaire Comparisons

Simulator Comparison P
Face validation questionnaire
    dVSS dVT .001
RM .316
    dVT dVSS .001
RM .092
    RM dVSS .316
dVT .001
Content validation questionnaire
    dVSS dVT .021
RM 1.00
    dVT dVSS .021
RM .092
    RM dVSS 1.00
dVT .092

The content validation–based questionnaire produced similar results. The dVSS was seen to have significant training utility compared with the dVT (P = .021). The RM was seen to again have a trend toward significance (P = .092) compared the dVT. There was no difference in utility between the dVSS and RM (P = 1.0) (Figure 2; Table 3).

Figure 2.

Figure 2.

Face and content total score comparisons.

The prices of the models ranged from $80,000 (for the da Vinci “backpack” alone) to $137,000 for the RM). Pricing also varies with the purchase of additional software with supplementary exercises for training (Figure 3).

Figure 3.

Figure 3.

Cost comparisons.

DISCUSSION

The use of minimally invasive surgical techniques continues to grow; however, the rapid expansion of the minimally invasive approach leads to a dilemma for surgical training on this equipment. The increase in robot-assisted and other forms of minimally invasive surgery has resulted in a greater need for training options for both residents and surgeons who did not receive this training during their residency. There have been ongoing attempts to create a standardized training program; however, training has not been standardized across the different virtual reality simulators. This study's goals were to provide a direct head-to-head comparison of the 3 most common commercially available simulators.

The dVT was one of the first VR simulators commercially available for robotic training and is still one of the most predominant. The console is simulated as a fully adjustable stereoscopic viewer and cable-driven master controller gimbals. The software operates on an external computer used for selecting tasks, managing users and curricula, and observing. All exercises were simulated by Mimic and were graded as an overall percentage composed of individual exercise-specific score modules. In comparison, the dVSS, commonly referred to as the backpack, is an optional add-on for da Vinci Xi and Si systems that is attached to the surgeon's console. Most of the exercises are simulated by Mimic and are also found on the dV- T, although there is a set of additional suturing exercises created by 3D Systems. The RM is a stand-alone simulated console, with nonfixed master controllers and all software developed by 3D Systems.

All 3 simulators offer a variety of VR tasks, ranging from basic orientation for console operation to procedure-specific skills. The dVT and RM have full-length procedures for common robotic cases, such as hysterectomy or prostatectomy. All of these stimulators have been independently validated for face and content validity.16,21,26 There have been comparisons between the dVT and dVSS, with the dVT showing slightly less overall performance; however, the difference was not directly quantified.29 In this study, we saw that, overall, all 3 simulators showed evidence of face and content validity, but that the dVSS scored significantly higher than its counterparts on both measures. This result is intuitive, given that it incorporates the da Vinci surgeon's console as part of the trainer itself. There was no significant difference between the other 2 trainers, although there was a trend toward a preference for the RM, for both its realism and usefulness for training. Although it may be intuitive that the dVSS would outperform the other 2 simulators, there has been no confirmation in the data up to this point. In addition, it is important to demonstrate that other simulators (such as the RM) are similar in performance to the dVSS.

The other factor evaluated in this study is the role of cost. Besides avoiding injury to the patient or possible liability during the known learning curve for robotic surgery, a motivation is the need to reduce cost during this period. It has been estimated in previous studies that the cost of the average learning curve in terms of operating time exceeds $200,000.30 This estimate does not include the price of the device itself or associated instruments and maintenance. It highlights the need for reliable simulators to help reduce this intraoperative learning curve, improving both patient safety and reducing cost.

The simulators themselves have various costs. The dVSS as a simulator is the least expensive at approximately $80,000; however this does not include the console itself (an additional $500,000 sunk cost), and the loss of OR time for use of the console (estimated at ∼$500/hour), unless an institution is willing to invest in a training-only console (which would decrease the price savings of the sunk cost of the console).21 However, as training is often scheduled after normal duty hours, there may not be a limiting factor for availability. The other trainers are similar in price at $110,000 for the dVT and $137,000 for the RM. Both the dVT and RM offer full-length procedural simulations, but the software still differs significantly. The dVT has Maestro AR technology available for partial nephrectomy, hysterectomy, and inguinal hernia repair, and both Xi- and Si-specific prostatectomy procedures. The choice-driven augmented reality interface is generated from virtual instruments overlaid onto intraoperative recordings. The RM modules include hysterectomy, prostatectomy, and lobectomy, each with segmented guidance.

Limitations of this study include the small sample size and the overall lesser experience of the cohort. The sample size limits the conclusions we can draw from the statistical trends, and it may be that, given a larger sample size, the trends established in the comparison of the RM and dVT would have been significant differences. This study, as opposed to previous studies, had a predominantly novice (defined variably in the literature based on previous robotic experience) experience level.16,21,26 Face and content validity are applicable to nonexperts; however, it limits the generalizability of the findings. This group was excellent for evaluating the simulators for the use of training. Simulators are being developed specifically to decrease the learning curve effect during surgery on actual patients.6 In general, trainees will use simulators more frequently than experienced surgeons. The assessment of face validity and usefulness of the simulator can ensure their use.

Future studies will benefit from evaluating skills acquisition and error rates between the simulators, as such data may significantly impact an institution's decision to acquire a certain simulator.

CONCLUSION

The 3 most commercially available VR robotic simulators all showed statistically significant evidence of face and content validity. Although the questionnaire scores for the dVSS outperformed the dVT and RM, the dVSS was limited by availability. The dVT and RM have similar cost and availability, with no difference in preference between them. The training modules available on the different models may be a differentiating feature between these 2 trainers, and both could be considered viable options for institutions seeking robotic training devices.

Contributor Information

Alexandria M. Hertz, Department of Urology.

Evalyn I. George, Department of Urology.

Christine M. Vaccaro, Department of Obstetrics and Gynecology, Madigan Army Medical Center, Tacoma, Washington, USA..

Timothy C. Brand, Department of Urology.

References:

  • 1. Barakat EE, Bedaiwy MA, Zimberg S, Nutter B, Nosseir M, Falcone T. Robotic-assisted, laparoscopic, and abdominal myomectomy: a comparison of surgical outcomes. Obstet Gynecol. 2011;117:256–266. [DOI] [PubMed] [Google Scholar]
  • 2. Tewari A, Srivasatava A, Menon M. A prospective comparison of radical retropubic and robot-assisted prostatectomy: experience in one institution. BJU Int. 2003;92:205–210. [DOI] [PubMed] [Google Scholar]
  • 3. Coelho RF, Rocco B, Patel MB. Retropubic, laparoscopic, and robot-assisted radical prostatectomy: a critical review of outcomes reported by high-volume centers. J Endourol. 2010;24:2003–2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4. Bell MC, Torgerson J, Seshadri-Kreaden U, Suttle AW, Hunt S. Comparison of outcomes and cost for endometrial cancer staging via traditional laparotomy, standard laparoscopy and robotic techniques. Gynecol Oncol. 2008;111:407–411. [DOI] [PubMed] [Google Scholar]
  • 5. Yu H-y, Hevelone ND, Lipsitz SR, Kowalczyk KJ, Hu JC. Use, costs and comparative effectiveness of robotic assisted, laparoscopic and open urological surgery. J Urol. 2012;187:1392–1399. [DOI] [PubMed] [Google Scholar]
  • 6. Patel VR, Palmer KJ, Coughlin G, Samavedi S. Robot-assisted laparoscopic radical prostatectomy: perioperative outcomes of 1500 cases. J Endourol. 2008;22:2299–2306. [DOI] [PubMed] [Google Scholar]
  • 7. Pradarelli JC, Campbell DA, Dimick JB. Hospital credentialing and privileging of surgeons: a potential safety blind spot. JAMA. 32015;13:1313–1314. [DOI] [PubMed] [Google Scholar]
  • 8. Satava RM. Historical review of surgical simulation: a personal perspective. World J Surg. 2008;32:141–148. [DOI] [PubMed] [Google Scholar]
  • 9. Fisher RA, Dasgupta P, Mottrie A. An over-view of robot assisted surgery curricula and the status of their validation. Int J Surg. 2015;13:115–123. [DOI] [PubMed] [Google Scholar]
  • 10. Culligan P, Gurshumov E, Lewis C, Priestly J, Komar J, Salamon C. Predictive validity of a training protocol using a robotic surgery simulator. Female Pelvic Med Reconstr Surg. 2014;20:48–51. [DOI] [PubMed] [Google Scholar]
  • 11. Dulan G, Rege RV, Hogg DC, Gilberg-Fisher KK, Tesfay ST, Scott DJ. Content and face validity of a comprehensive robotic skills training program for general surgery, urology, and gynecology. Am J Surg. 2012;203:535–539. [DOI] [PubMed] [Google Scholar]
  • 12. Smith R, Patel V, Satava R. Fundamentals of robotic surgery: a course of basic robotic surgery skills based upon a 14-society consensus template of outcomes measures and curriculum development. Int J Med Robot. 2014;10:379–384. [DOI] [PubMed] [Google Scholar]
  • 13. Satava R, Smith R, Patel V. Fundamentals of robotic surgery: outcomes measures and curriculum development. Int J Med Robot. 2014;10;379–384. [DOI] [PubMed] [Google Scholar]
  • 14. Tanaka A, Graddy C, Simpson K, Perez M, Truong M, Smith R. Robotic surgery simulation validity and usability comparative analysis. Surg Endosc. 2016;30:3720–3729. [DOI] [PubMed] [Google Scholar]
  • 15. Abboudi H, Khan MS, Aboumarzouk O, et al. Current status of validation for robotic surgery simulators: a systematic review. BJU Int. 2013;111:194–205. [DOI] [PubMed] [Google Scholar]
  • 16. Hung AJ, Zehnder P, Patil MB, et al. Face, content and construct validity of a novel robotic surgery simulator. J Urol. 2011;186:1019–1024. [DOI] [PubMed] [Google Scholar]
  • 17. Kelly DC, Margules KC, Kundavaram K, et al. Face, content, and construct validation of the da Vinci Skills Simulator. Urology. 2012;79:1068–1072. [DOI] [PubMed] [Google Scholar]
  • 18. Liss MA, Abdelshehid C, Quach S, et al. Validation, correlation, and comparison of the da Vinci trainer™ and the da Vinci surgical skills simulator™ using the Mimic™ software for urologic robotic surgical education. J Endourol. 2012;26:1629–1634. [DOI] [PubMed] [Google Scholar]
  • 19. Kenney PA, Wszolek MF, Gould JJ, Libertino JA, Moinzadeh A. Face, content, and construct validity of dV-trainer, a novel virtual reality simulator for robotic surgery. Urology. 2009;73:1288–1292. [DOI] [PubMed] [Google Scholar]
  • 20. Sethi AS, Peine WJ, Mohammadi Y, Sundaram CP. Validation of a novel virtual reality robotic simulator. J Endourol. 2009;23:503–508. [DOI] [PubMed] [Google Scholar]
  • 21. Perrenot C, Perez M, Tran N, Jehl JP, Felblinger J, Bresler L. The virtual reality simulator dV-Trainer(R) is a valid assessment tool for robotic surgical skills. Surg Endosc. 262;587–593, 2012. [DOI] [PubMed] [Google Scholar]
  • 22. Korets R, Gaverson JA, Mues AC, et al. Face and construct validity assessment of 2nd generation robotic surgery simulator. J Urol. 2011;185:e488. [Google Scholar]
  • 23. Korets R, Mues AC, Graversen JA, et al. Validating the use of the Mimic dV-trainer for robotic surgery skill acquisition among urology residents. Urology. 2011;78;1326–1330. [DOI] [PubMed] [Google Scholar]
  • 24. Lee JY, Mucksavage P, Kerbl DC, Huynh VB, Etafy M, McDougall EM. Validation study of a virtual reality robotic simulator: role as an assessment tool? J Urol. 2012;187;998–1002. [DOI] [PubMed] [Google Scholar]
  • 25. Finnegan KT, Meraney AM, Staff I, Shichman SJ. da Vinci Skills Simulator construct validation study: correlation of prior robotic experience with overall score and time score simulator performance. Urology. 2012;80;330–335. [DOI] [PubMed] [Google Scholar]
  • 26. Whittaker G, Aydin A, Raison N, et al. Validation of the RobotiX Mentor robotic surgery simulator. J Endourol. 2016;30:338–346. [DOI] [PubMed] [Google Scholar]
  • 27. Lallas CD, Davis JW, Surgeons Members Of The Society Of Urologic Robotic. Robotic surgery training with commercially available simulation systems in 2011: a current review and practice pattern survey from the society of urologic robotic surgeons. J Endourol. 2012;26;283–293. [DOI] [PubMed] [Google Scholar]
  • 28. McDonough PS, Tausch TJ, Peterson AC, Brand TC. Initial validation of the ProMIS surgical simulator as an objective measure of robotic task performance. J Robot Surg. 2011;5:195–199. [DOI] [PubMed] [Google Scholar]
  • 29. Liss MA, Tausch TJ, Peterson AC, Brand TC. Validation, correlation, and comparison of the da Vinci trainer(TM) and the daVinci surgical skills simulator(TM) using the Mimic(TM) software for urologic robotic surgical education. J Endourol. 2012;26;1629–1634. [DOI] [PubMed] [Google Scholar]
  • 30. Steinberg PL, Merguerian PA, Bihrle W, 3rd, Seigne JD. The cost of learning robotic-assisted prostatectomy. Urology. 2008;72;1068–1072. [DOI] [PubMed] [Google Scholar]

Articles from JSLS : Journal of the Society of Laparoendoscopic Surgeons are provided here courtesy of Society of Laparoscopic & Robotic Surgeons

RESOURCES