Skip to main content
Frontiers in Psychology logoLink to Frontiers in Psychology
. 2020 Sep 10;11:542752. doi: 10.3389/fpsyg.2020.542752

Distinguishing Intermediate and Novice Surgeons by Eye Movements

Gonca Gokce Menekse Dalveren 1,2,*, Nergiz Ercil Cagiltay 3
PMCID: PMC7511664  PMID: 33013592

Abstract

Surgical skill-level assessment is key to collecting the required feedback and adapting the educational programs accordingly. Currently, these assessments for the minimal invasive surgery programs are primarily based on subjective methods, and there is no consensus on skill level classifications. One of the most detailed of these classifications categorize skill levels as beginner, novice, intermediate, sub-expert, and expert. To properly integrate skill assessment into minimal invasive surgical education programs and provide skill-based training alternatives, it is necessary to classify the skill levels in as detailed a way as possible and identify the differences between all skill levels in an objective manner. Yet, despite the existence of very encouraging results in the literature, most of the studies have been conducted to better understand the differences between novice and expert surgical skill levels leaving out the other crucial skill levels between them. Additionally, there are very limited studies by considering the eye-movement behaviors of surgical residents. To this end, the present study attempted to distinguish novice- and intermediate-level surgical residents based on their eye movements. The eye-movement data was recorded from 23 volunteer surgical residents while they were performing four computer-based simulated surgical tasks under different hand conditions. The data was analyzed using logistic regression to estimate the skill levels of both groups. The best results of the estimation revealing a 91.3% recognition rate of predicting novice and intermediate surgical residents on one scenario were selected from four under the dominant hand condition. These results show that the eye-movements can be potentially used to identify surgeons with intermediate and novice skills. However, the results also indicate that the order in which the scenarios are provided, and the design of the scenario, the tasks, and their appropriateness with the skill levels of the participants are all critical factors to be considered in improving the estimation ratio, and hence require thorough assessment for future research.

Keywords: eye movement events, eye movement classification, eye tracking, surgical skill assessment, surgical training

Introduction

Surgical skill-level assessment is a critical process to ensure competence and prevent clinical errors while developing effective instructional methods for training. It is commonly acknowledged that the skill levels of surgeons can vary depending on numerous elements (Uhrich et al., 2002; Silvennoinen et al., 2009; Chandra et al., 2010; De Blacam et al., 2012). Currently, the assessment of surgical residents’ skill levels is based on subjective approaches and conventional methods, and as such, are prone to bias and questionable rationality (Resnick et al., 1991; Wanzel et al., 2002; Richstone et al., 2010; Meneghetti et al., 2012). Hence, the traditional assessment methods have many shortcomings, such as subjective evaluation, need for expert surgeons and requiring standardized methods of assessment (Moorthy et al., 2003; Ahmidi et al., 2012; Zheng et al., 2015). Several studies have reported that more effective skill-level assessment techniques and training evaluation systems could improve skill-based training, and accordingly patient healthcare (Reiley et al., 2011; D’Angelo et al., 2015; Zheng et al., 2015). Such assessment is also necessary to adapt the content and level of training programs to the surgical residents’ needs and experience, providing sufficient feedback for both trainees and educators, and improving the curriculum (De Blacam et al., 2012; Cundy et al., 2015; Cagiltay et al., 2017; Mizota et al., 2020). On the other hand, simulation environments allow trainees to develop knowledge and necessary skills in a safe and effective environment to apply in real-world applications (Earle, 2006; Van Sickle et al., 2006; Carter et al., 2010; Cristancho et al., 2012; Bilgic et al., 2019). Thus, these virtual and interactive settings in computer-based systems are risk-free (Gallagher and Satava, 2002; Maran and Glavin, 2003; Earle, 2006; Van Sickle et al., 2006; Loukas et al., 2011; Seagull, 2012). Accordingly, several training scenarios have been developed for surgical-skill improvement and assessment requirements for surgery purposes (McNatt and Smith, 2001; Gallagher and Satava, 2002; MacDonald et al., 2003; Jones, 2007; Curtis et al., 2012; Feldman et al., 2012; Hsu et al., 2015). Oropesa et al. (2011) report that adaptation of newly developed tools and techniques to surgical training programs has the possibility to realize an essential role for the assessment of surgical skills. However, there are currently very limited studies considering the precise assessment of these skill levels and they have mainly used the performance metrics for the assessment purposes. For instance, in previous studies, surgeons’ experience levels were assessed for each task in accordance to time, errors, and economy of movement for each hand (McNatt and Smith, 2001; Adrales et al., 2003; Arora et al., 2010; Meneghetti et al., 2012; Perrenot et al., 2012; Mathis et al., 2020). An earlier study objectively estimated the skill levels in a virtual surgical training environment using performance data from surgical residents, such as task completion time, distance, and success (Topalli and Cagiltay, 2019). Another study used electromyography (EMG) activity and muscular discomfort scores to assess the surgical skill levels of surgeons (Uhrich et al., 2002). Oostema et al. (2008) quantified the laparoscopic operative experience for each user using time, path length, and smoothness. Similarly, McDougall et al. (2006) attempted to distinguish surgeons who were laparoscopically naive and those with laparoscopic experience regardless of the degree of previous related surgical experience (McDougall et al., 2006). Oropesa et al. (2013) used motion metrics for assessing laparoscopic psychomotor skills from video analysis.

New technologies offer advantages through recording the eye movements (Atkins et al., 2013; Zheng et al., 2015; Kruger and Doherty, 2016; Gegenfurtner et al., 2017; Jian and Ko, 2017) of surgeons and analyzing the obtained data to provide a cost-effective, automated, and objective basis for assessing their skill levels (Ahmidi et al., 2012; Tien et al., 2014). In this respect, eye-tracking provides objective metrics about human behavior (Yarbus, 1967; Gegenfurtner and Seppänen, 2013; Dogusoy-Taylan and Cagiltay, 2014; Piccardi et al., 2016; Tsai et al., 2016; Bröhl et al., 2017; Jian and Ko, 2017; Tricoche et al., 2020), and these systems have many beneficial properties, making it easy to record and analyze eye-movement data (Gegenfurtner and Seppänen, 2013; Tien et al., 2015; Zheng et al., 2015; Kruger and Doherty, 2016; Gegenfurtner et al., 2017). Thus, eye-tracking is used for assessing and understanding the differences between skill levels in the medical domain (Stuijfzand et al., 2016; Gegenfurtner et al., 2017; McLaughlin et al., 2017; Fichtel et al., 2019). The differences in performances are related to the information-processing capabilities of the left and right hemispheres of the brain, implying that when visual control is required, the dominant hand is likely to perform better than the non-dominant hand and both hands (Hoffmann, 1997). For instance, Vickers (1995) reported the difference in eye movements between expert and novice sports players in foul shooting, in which expert players were seen not to follow the whole shooting process with their eyes but novice players used their eyes to adjust their shots before releasing the ball. In another different though related study, Kasarskis et al. (2001) investigated the differences between the eye movements of expert and novice pilots in a landing simulation and showed that the fixation times of novices were longer than those of the experts because the latter assembled the necessary information more rapidly than the former. Furthermore, the study of Khan et al. (2012) reports significant differences between surgeons from different skill levels on their gaze patterns while watching surgical videos. Another study shows that according to the results of a meta-analysis fixation duration of non-experts is longer than the experts (Gegenfurtner et al., 2011). Understanding these differences, classifying different skill-leveled surgeons appropriately and objectively is another critical task for surgical education programs (McNatt and Smith, 2001; Uhrich et al., 2002; Perrenot et al., 2012; Doleck et al., 2016). By objectively classifying the experience levels of participants, adaptive systems can be developed for their training (Perrenot et al., 2012), and objective standards can be generated (Oostema et al., 2008) as an evaluation criterion for student admission to such programs. However, in the literature, there is no study reporting the precise estimations of the surgeons’ skill levels, especially in relation to their endoscopic surgical skills and eye movement behaviors.

Additionally, there is no consensus on the classification of the skill levels of minimal invasive surgery procedures. Some studies classify it as novice and expert (Aggarwal et al., 2006; Fichtel et al., 2019), where others consider it as novice, intermediate, and expert (Grantcharov et al., 2003; Schreuder et al., 2009; Shetty et al., 2012). Silvennoinen et al. (2009) presented the most detailed classification of the expertise levels of minimally invasive surgery in the following five stages: a beginner who has simply non-specialist knowledge, a novice who has started to develop initial knowledge in the domain, an intermediate who has intensified his/her knowledge, a sub-expert who has general knowledge but incomplete knowledge of a specialized domain, and an expert who has deep knowledge of a specialized domain (Silvennoinen et al., 2009). Earlier studies were mainly conducted with novices and experts for determining their skill levels (Vickers, 1995; Rosser et al., 1998; Kasarskis et al., 2001; Richstone et al., 2010). Even though there have been other efforts to classify the experience levels of surgeons, these were mostly to differentiate novices and experts using a box trainer from their hand movements (Ahmidi et al., 2010; Horeman et al., 2011, 2013; van Empel et al., 2012). Hence, there are very few studies classifying the novice and intermediate endoscopic surgical skill levels based on eye movements with one earlier study (Chmarra et al., 2010) reporting 74.2% accuracy. Although there are several studies referring to simulation based training environments for endoscopic surgery procedures, they are very limited in classifying the endoscopic surgery skill levels in a detail manner and at a higher level of accuracy. Furthermore, there are very limited studies that focus on better understanding the possible contributions of eye-movement analyses on these assessment procedures.

The main aim of this study was to provide insights into novice and intermediate surgical skill level assessment techniques using eye-tacking technology in virtual reality settings. First, it attempts to show a possible relationship between the eye-movements and surgical skill levels for intermediates and novices. Afterward, it attempts to classify these skill levels through their eye-movement behaviors. The purpose behind this initiative was to integrate skill assessment into the curriculum of surgical education programs to provide appropriate feedback and understand the differences between these skill levels. In particular, compared to the gap between novice and expert groups being higher (Silvennoinen et al., 2009), differentiating novice and intermediate skill level groups is particularly challenging. Additionally, to the best of our knowledge, no study has been found that classified these groups based on eye-movement events. For instance, the study of Khan et al. (2012) reports differences on gaze patterns of surgical residents while watching surgical videos. According to the results of this study it was reported that novice surgeons’ eye-gaze patterns generally roam from key areas of the operational field, whereas expert surgeons’ eye-gaze patterns focused on the task-relevant areas (Khan et al., 2012). Also, the study of Gegenfurtner et al. (2011) shows differences on fixation durations between experts and non-experts. According to this study, experts had shorter fixation durations compared to the non-experts (Gegenfurtner et al., 2011). The results of this earlier studies provide evidences indicating differences between novice and intermediate surgical residences’ eye movements while performing surgical tasks. However, these studies consider the differences between experts and non-experts or novices. Hence, earlier studies are very limited considering eye movement behavioral differences of novice and intermediate groups. Therefore, this study was conducted with novices and intermediates and the main hypothesis of the study is that there is a significant difference between intermediate and novice surgical residents’ eye movements considering number of fixations, fixation durations and number of saccades while they are performing surgical tasks under different hand conditions (dominant-hand, non-dominant hand, and both hand). The second hypothesis is that the novice and intermediate surgical residents can be classified by using their eye movements considering number of fixations while they are performing surgical tasks under different hand conditions (dominant-hand, non-dominant hand, and both hand). Hence, the main contribution of this study is classifying the novice- and intermediate-group surgeons through their eye movement events to predict their skill levels accordingly.

For this purpose, in this study, four surgical scenarios were applied and performed in a virtual reality environment under different hand conditions (dominant hand, non-dominant hand, and both hands) using haptic devices controlled by surgical residents. Their eye-movement data was collected while they performed the given tasks in these scenarios. The data was classified into number of fixations, fixation duration and number of saccades events with an open-source eye-movement classification algorithm, Binocular Individual Threshold (BIT) (van der Lans et al., 2011), a velocity-based algorithm for identifying fixations and saccades. BIT is effective since it is a parameter-free fixation-identification mechanism which automatically identifies task- and individual-specific velocity thresholds (van der Lans et al., 2011; Menekse Dalveren and Cagiltay, 2019). BIT classifies fixations and saccades from the data of both eyes based on individual-specific thresholds. This algorithm has advantages over other similar algorithms such as containing binocular viewing and being independent of machine and sampling frequency (van der Lans et al., 2011). Therefore, BIT algorithm can classify the data from varying eye-tracking devices with various sampling frequencies and different level of sensitivities (van der Lans et al., 2011).

Materials and Methods

In this study, the eye movements of surgical residents were investigated to distinguish their surgical skill levels. Recorded raw eye data was classified and analyzed for differentiating the intermediate and novice surgical residents. This study was carried out in accordance with the protocol which was approved by the Human Research Ethics Board of Atılım University. All subjects gave verbal and written informed consent in accordance with the Declaration of Helsinki.

Participants

Twenty-three volunteer participants from two departments were recruited for this study: 13 from neurosurgery and 10 from ear-nose-throat (ENT) surgery. Since it is difficult to access surgeons in specialized fields to volunteer for such initiatives, this number of participants is usually considered as acceptable. The majority of the earlier studies were also conducted with limited number of participants, such as 14 surgeons (Wilson et al., 2010), 22 surgeons (Cope et al., 2015), 9 neurosurgeons (Eivazi et al., 2017), 12 students (Uemura et al., 2016), and 19 surgical residents (Hsu et al., 2015). Of the 23 participants in the current study, 14 (three females) were novices with an average age of 27.71 (SD = 6.96), with no one having previously performed an endoscopic surgery on their own. On average, the 14 participants had observed 9.57 (SD = 13.51) and assisted in 3.57 (SD = 10.64) surgical procedures. The remaining nine were intermediates with an average age of 29.33 (SD = 1.50) years, who had observed 48.33 (SD = 31.62) and assisted in 32.00 (SD = 24.19) surgical operations, while performing 16.56 (SD = 16.60) operations on average on their own.

Simulated Scenarios

The simulation scenarios were developed to address the development of required skills for endoscopic skull-base surgery operations which are mainly conducted by ENT and neurosurgery experts. Skills, such as depth perception, 2D and 3D environmental transformations, left-right hand coordination, and hand-eye coordination were aimed to be addressed through the designed scenarios, which were based on surgical experts’ guidance from the neurosurgery and ENT departments of the medical school.

Virtually simulated environments resemble the real world with their visualization and interaction properties (Perrenot et al., 2012; Zhang et al., 2017). In this study, four scenarios were developed based on surgical skill development requirements for endoscopic surgery purposes to collect the participants’ eye movement data.

Scenario 1

In this scenario, the participants are expected to use an operational instrument through a haptic device. The scenario is designed to improve depth perception, camera control, 2D-3D conversion, and effective tool usage. With the aid of this tool, the participants have to catch a red ball, which appears randomly in different locations in a room (Figure 1A). The ball must be placed close to the cube; then, the color of the ball turns green. The participants must grab this ball and place it on the green cube. The cube also appears at random positions in the room. This task is repeated 10 times.

FIGURE 1.

FIGURE 1

Scenarios [(A) scenario 1, (B) scenario 2, (C) scenario 3, and (D) scenario 4].

Scenario 2

A red ball appears randomly in one of the boxes as shown in Figure 1B. The red ball will explode if the approach is at the right angle using the haptic device. This process is repeated 10 times with the aim of developing depth perception and improving the participant’s ability to approach a certain point from the correct angle.

Scenario 3

This scenario is a simulated surgical model, in which the participant feels as if they were actually undertaking surgery. As depicted in Figure 1C, the model shows the inside of a human nose. The participants are expected to remove the 10 objects located at different places within the model using the surgical tool. They can move the tool within the model using the haptic device and feel the tissue as the device gives force feedback upon collision with any surface. By using the surgical tool in the most accurate way the participant can complete the operation by carefully removing all the objects from the model.

Scenario 4

This scenario is a simulated surgical model and designed like inside of a human nose with similar texture, simulating the field vision of a surgeon during an actual operation. The main task in this scenario is to move an object (the white ball) smoothly on a specified path (Figure 1D). The task is completed when the participant reaches the end of the specified path. The important aspect in this scenario is using the haptic device at a proper angle to keep the object moving on the path.

Procedure

The participants were informed about the purpose, procedures, and duration of the experiment. The eye-movement data of the surgical residents were collected using an eye-tracking device while they performed the surgical tasks. For the eye data recordings, the Eye Tribe tracker was positioned under the monitor (Figure 2) to provide the information regarding the distance of the user to the monitor and the visibility of the eyes. Therefore, the participants were asked to adjust their seating position to a distance of 70 cm in front of the tracker. The recording frequency was set to 60 Hz and a 9-point calibration process was performed.

FIGURE 2.

FIGURE 2

Experiment design.

The scenarios were ordered from easy to difficult levels with surgical experts’ guidance from the neurosurgery and ENT departments of the medical school. All the participants performed the scenarios in the same order of 1, 2, 3, and 4 to reduce the chance of negative effects on learning. Each scenario has 10 tasks with a fixed period of time as 20 s. Therefore, each scenario takes 200 s (20 × 10). For added objectivity and reduced order effect, 12 of the participants performed the tasks using their dominant hand, and the remainder with their non-dominant hand. In all cases, the tasks were last performed under the both-hands condition. Each scenario was performed by surgeons using Geomagic Touch haptic devices. All scenarios took place in a simulated environment using a surgical instrument under virtually lit conditions in dominant and non-dominant hand conditions, whereas in the both-hands condition, the participants had to use the light source and the surgical instrument in a coordinated fashion. Therefore, in the both-hands condition, the participants were required to use two haptic devices. They used the first haptic device to hold the surgical instrument with their dominant hand, and they held the second haptic device with their non-dominant hand as a light source to illuminate the operating area as shown in Figure 3. Since the Eye Tribe eye tracker can only record the eye movement data in x and y coordinates, there is a need to classify this raw data into eye movement events, such as fixations; a slow-period eye movement with reduced dispersion and velocity. The classification of raw eye data was performed with BIT, which is an open-source eye movement classification algorithm capable of categorizing raw eye data into fixations and saccades independently from the eye tracker and without specifying the threshold values (van der Lans et al., 2011). In addition, the classified eye movement data was analyzed using statistical methods to predict the skill levels of surgeons based on their eye-movement behaviors.

FIGURE 3.

FIGURE 3

Both-hands condition.

Analysis

The analysis of all the data classified by the BIT algorithm was performed using SPSS for the Windows software package (version 23; IBM Corporation, New York, United States) at 95% confidence level. Since the sample size was 23 and the normality assumptions were violated, the non-parametric test technique of Mann-Whitney was used (McCrum-Gardner, 2008) to identify differences between the intermediate and novice surgeons in terms of eye-movement data. This technique is used to compare two independent groups in relation to a quantitative variable (McCrum-Gardner, 2008). A logistic regression analysis was conducted to predict the skill levels of the surgical residents.

Results

According to the Mann-Whitney test for the novices and intermediates, there was a statistically significant difference in the number of fixations between the two groups (Figure 4) for Scenario 1 under the dominant hand condition. In other scenarios and conditions, no significant difference was found for number of fixations, fixation durations and number of saccades. The scatter plot for number of fixations of Scenario 1 under dominant hand condition was given in Figure 5. In Scenario 1, under the dominant hand condition, the novice surgeons fixated more than the intermediate surgeons according to the results of the BIT algorithm (U = 16, p < 0.05, z = −2.980, r = −0.62). This result reveals that there is a potential difference between the novice and intermediate surgical residents with the experience level having an impact on the number of fixations and based on the r value (−0.62) the effect size shows that the difference between intermediate and novice surgeons can be considered as high (Rosenthal et al., 1994). Also, another statistical power analysis was performed to evaluate the effect sizes based on data of this study and the Eta squared (η2) and d Cohen values were found as 0.38 and 1.57, respectively (Lenhard and Lenhard, 2016).

FIGURE 4.

FIGURE 4

Number of fixations of novice and intermediate surgeons.

FIGURE 5.

FIGURE 5

Scatter of the data for number of fixations.

The logistic regression analysis ascertained the effects of the number of fixations, fixation duration and the dominant hand on the prediction of the participants’ skill level. According to the results, in the dominant-hand condition for Scenario 1, the logistic regression model was statistically significant, x2(3) = 15.661, p = 0.001. The model explained 66.9% (Negelkerke R2) of the variance in the skill level, and correctly classified 91.3% of all the cases. For Scenario 2, the logistic regression model was not statistically significant, x2(3) = 6.445, p = 0.092. The model explained 33.1% (Negelkerke R2) of the variance in the skill level, while accurately classifying 69.6% of the cases. For Scenario 3, the logistic regression model was not statistically significant, x2(3) = 5.449, p = 0.142. The model explained 28.6% (Negelkerke R2) of the variance in the skill level with 65.2% of the cases correctly classified. Lastly, in Scenario 4, the model was not statistically significant, x2(3) = 4.941, p = 0.176, explaining 26.2% (Negelkerke R2) of the variance in the skill level and correctly classifying 56.5% of the cases.

According to the results, in the non-dominant hand condition for Scenario 1, the logistic regression model was statistically significant, x2(3) = 9.605, p = 0.022. The model explained 46.3% (Negelkerke R2) of the variance in the skill level and correctly classified 65.2% of all cases. For Scenario 2, the model was also not statistically significant, x2(3) = 4.830, p = 0.185, explaining 25.7% (Negelkerke R2) of the variance in the skill level and correctly classifying 56.5% of the cases. For Scenario 3, the logistic regression model was statistically significant, x2(3) = 10.375, p = 0.016, explaining 49.2% (Negelkerke R2) of the variance in the skill level and correctly classifying 73.9% of all cases. For Scenario 4, the model was not statistically significant, x2(3) = 5.388, p = 0.146 since it explained 28.3% (Negelkerke R2) of the variance in the skill level and correctly classified 56.5% of the cases.

Finally, in the both-hand condition for Scenario 1 the logistic regression model was statistically significant, x2(3) = 14.289, p = 0.003, explaining 62.7% (Negelkerke R2) of the variance in the skill level and correctly classifying 73.9% of all cases. For Scenario 2, the model was not statistically significant, x2(3) = 7.440, p = 0.059 as it 37.5% (Negelkerke R2) of the variance and classified 65.2% of the cases. For Scenario 3, the model was statistically significant, x2(3) = 8.670, p = 0.034 and explained 42.6% (Negelkerke R2) of the variance in the skill level and correctly classified 65.2% of all cases. For Scenario 4, the logistic regression model was again not statistically significant, x2(3) = 8.031, p = 0.45 as it explained 39.9% (Negelkerke R2) of the variance in the skill level and correctly classified 60.9% of the cases.

Discussion

The results of this study show that it is possible to objectively evaluate the skill levels of novice and intermediate surgeons using eye-movement events, such as number of fixations. This result supports the findings of earlier studies (Gallagher and Satava, 2002; Andreatta et al., 2008; Richstone et al., 2010; Gegenfurtner et al., 2011; Khan et al., 2012). For instance, studies indicating that surgical skill levels can be objectively evaluated by eye-tracking metrics through virtually simulated and live environments (Andreatta et al., 2008; Richstone et al., 2010). Additionally, the findings are in parallel with the results of another earlier study showing that experienced laparoscopic surgeons performed the tasks significantly faster, with fewer errors and more economy in the movement of instruments, and greater consistency in performance (Gallagher and Satava, 2002). Although this is more challenging than comparatively evaluating the skill levels of novice and expert surgeons, this study shows that it is possible to classify intermediate and novice level surgical residents through their eye movement behaviors with a high accuracy. Hence, the main contribution of this current study is showing the eye movement behavioral differences of novice and intermediate surgeons and classifying these groups with a high accuracy (91.3%) which are not reported in the earlier studies. This assessment process could be used to accurately monitor the acquisition of skills throughout training programs at a much earlier stage.

However, the results indicate that in different scenarios, the accuracy of estimations may change (Table 1). Iqbal et al. (2004) reported that the difficulty level of the tasks demanded longer processing times and brought higher subjective ratings of mental workload. In other words, the difficulty level of scenarios needs to be designed in parallel with the skill levels of the trainees. When the tasks in the scenario are too easy or too hard in relation to the trainees’ skill levels, then it could be hard to assess their levels with these tasks. Another study stated that the fidelity level was an important factor affecting mental workload (Munshi et al., 2015). Accordingly, the fidelity levels of the scenarios could be another consideration in designing training scenarios according to a specific skill level. For a better integration of these simulation-based training environments into traditional surgical education programs, further studies need to be conducted to standardize these training scenarios. Therefore, the order of the scenarios and the design of the tasks defined for each scenario may also have an effect on these different findings, necessitating additional research to gain a more thorough understanding of the affective factors on the accuracy of estimations.

TABLE 1.

Logistic regression results for all conditions and scenarios.

Scenario Dominant hand
Non-dominant hand
Both hand
% p % p % p
1 91.3 0.001* 65.2 0.020* 73.9 0.003*
2 69.6 0.092 56.6 0.185 65.2 0.059
3 65.2 0.142 73.9 0.016* 65.2 0.034*
4 56.5 0.176 56.5 0.146 60.9 0.450

*Significant results (p < 0.05).

Conclusion

The use of eye-movement events is an objective approach which does not require the time and expenditure required to employ expert evaluators. The results of this study reveal that by appropriately designing scenarios addressing both skills under specific conditions (as in the case of the dominant hand condition in Scenario 1 in this study) it may be possible to distinguish the levels of intermediate and novice surgeons through their eye behaviors, up to 91.3%. The highest ratio of the estimation was achieved in Scenario 1 under the dominant hand condition, which implies several possibilities. First, the content and design of the assessment scenarios should be in line with the skill levels of the participants. Second, the design of the scenarios and tasks that require specific skills and their appropriateness with the skill levels of the participants is another critical factor in improving the estimation ratio. In other words, for this group of participants, the design of Scenario 1 was found to be more appropriate for skill assessment purposes. Third, the order of the scenarios implemented in the research procedure is also of significance; for example, as Scenario 1 was implemented first, the accuracy of the assessment regarding the other scenarios might have declined due to the learning affect.

In our earlier study (Topalli and Cagiltay, 2019), the best estimation accuracy was 86% using features, such as time, distance, camera distance, success, catch time, error time, error distance, and deviation count under dominant hand, non-dominant hand and both-hands conditions. This study showed better accuracy when the participants’ eye-movement events were analyzed. Accordingly, this study provides evidence that the eye-movement behaviors of the surgical residents are very important for assessing their skill levels.

Based on the results of this study, eye movements can provide insights into the skill levels of surgical residents and their current education program. In light of this information, it might be beneficial to properly design and organize simulation-based instructional systems for surgical training programs integrating eye movement behavioral analysis into their current methods. Additionally, the progress levels of trainees can be continuously and objectively assessed, and accordingly guidance can be given during training programs.

In conclusion, performance evaluations through computer-based simulation environments by considering the eye movements of surgeons can be improved. In the near future, data obtained from the eye movements of surgical residents can potentially positively support the current evaluation methods by adding to the degree of objectivity. Future studies considering the eye movement data of surgical residents synchronized with their performance data, as well as hand movement data can provide even better results on these estimations and potentially improve the level of prediction. Additionally, by analyzing the performance data (such as task accuracy, task duration and number of errors) in correlation with the eye movements of the trainees’ deeper insights about the possible relationships between eye movements and surgical skill levels can be gained. Furthermore, future studies are required in consideration of the higher level of expertise in the endoscopic surgery field.

Data Availability Statement

The datasets generated for this study are available on request to the corresponding author.

Ethics Statement

The studies involving human participants were reviewed and approved by the Ethics Committee of the Atılım University. The participants provided their written informed consent to participate in this study.

Author Contributions

NC designed the experiment. GM analyzed the data. GM and NC wrote the article. Both authors contributed to the article and approved the submitted version.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Acknowledgments

We would like to thank the Hacettepe University Medical School for their valuable support throughout the research.

References

  1. Adrales G. L., Park A. E., Chu U. B., Witzke D. B., Donnelly M. B., Hoskins J. D., et al. (2003). A valid method of laparoscopic simulation training and competence assessment1, 2. J. Surg. Res. 114 156–162. 10.1016/s0022-4804(03)00315-9 [DOI] [PubMed] [Google Scholar]
  2. Aggarwal R., Grantcharov T., Moorthy K., Hance J., Darzi A. (2006). A competency-based virtual reality training curriculum for the acquisition of laparoscopic psychomotor skill. Am. J. Surg. 191 128–133. 10.1016/j.amjsurg.2005.10.014 [DOI] [PubMed] [Google Scholar]
  3. Ahmidi N., Hager G. D., Ishii L., Fichtinger G., Gallia G. L., Ishii M. (2012). “Surgical task and skill classification from eye tracking and tool motion in minimally invasive surgery,” in Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, eds Jiang T., Navab N., Pluim J. P. W., Viergever M. A. (Berlin: Springer; ), 295–302. 10.1007/978-3-642-15711-0_37 [DOI] [PubMed] [Google Scholar]
  4. Ahmidi N., Ishii M., Fichtinger G., Gallia G. L., Hager G. D. (2010). An objective and automated method for assessing surgical skill in endoscopic sinus surgery using eye-tracking and tool-motion data. Int. Forum Allergy Rhinol. 2 507–515. 10.1002/alr.21053 [DOI] [PubMed] [Google Scholar]
  5. Andreatta P. B., Woodrum D. T., Gauger P. G., Minter R. M. (2008). LapMentor metrics possess limited construct validity. Simul. Healthc. 3 16–25. 10.1097/sih.0b013e31816366b9 [DOI] [PubMed] [Google Scholar]
  6. Arora S., Sevdalis N., Nestel D., Woloshynowych M., Darzi A., Kneebone R. (2010). The impact of stress on surgical performance: a systematic review of the literature. Surgery 147 318–330, 330.e1–330.e6. [DOI] [PubMed] [Google Scholar]
  7. Atkins M. S., Tien G., Khan R. S., Meneghetti A., Zheng B. (2013). What do surgeons see: capturing and synchronizing eye gaze for surgery applications. Surg. Innov. 20 241–248. 10.1177/1553350612449075 [DOI] [PubMed] [Google Scholar]
  8. Bilgic E., Alyafi M., Hada T., Landry T., Fried G. M., Vassiliou M. C. (2019). Simulation platforms to assess laparoscopic suturing skills: a scoping review. Surg. Endosc. 33 2742–2762. 10.1007/s00464-019-06821-y [DOI] [PubMed] [Google Scholar]
  9. Bröhl C., Theis S., Rasche P., Wille M., Mertens A., Schlick C. M. (2017). Neuroergonomic analysis of perihand space: effects of hand proximity on eye-tracking measures and performance in a visual search task. Behav. Inf. Technol. 36 737–744. 10.1080/0144929x.2016.1278561 [DOI] [Google Scholar]
  10. Cagiltay N. E., Ozcelik E., Sengul G., Berker M. (2017). Construct and face validity of the educational computer-based environment (ECE) assessment scenarios for basic endoneurosurgery skills. Surg. Endosc. 31 4485–4495. 10.1007/s00464-017-5502-4 [DOI] [PubMed] [Google Scholar]
  11. Carter Y. M., Wilson B. M., Hall E., Marshall M. B. (2010). Multipurpose simulator for technical skill development in thoracic surgery. J. Surg. Res. 163 186–191. 10.1016/j.jss.2010.04.051 [DOI] [PubMed] [Google Scholar]
  12. Chandra V., Nehra D., Parent R., Woo R., Reyes R., Hernandez-Boussard T., et al. (2010). A comparison of laparoscopic and robotic assisted suturing performance by experts and novices. Surgery 147 830–839. 10.1016/j.surg.2009.11.002 [DOI] [PubMed] [Google Scholar]
  13. Chmarra M. K., Klein S., De Winter J. C., Jansen F.-W., Dankelman J. (2010). Objective classification of residents based on their psychomotor laparoscopic skills. Surg. Endosc. 24 1031–1039. 10.1007/s00464-009-0721-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Cope A. C., Mavroveli S., Bezemer J., Hanna G. B., Kneebone R. (2015). Making meaning from sensory cues: a qualitative investigation of postgraduate learning in the operating room. Acad. Med. 90 1125–1131. 10.1097/acm.0000000000000740 [DOI] [PubMed] [Google Scholar]
  15. Cristancho S., Moussa F., Dubrowski A. (2012). Simulation-augmented training program for off-pump coronary artery bypass surgery: developing and validating performance assessments. Surgery 151 785–795. 10.1016/j.surg.2012.03.015 [DOI] [PubMed] [Google Scholar]
  16. Cundy T. P., Thangaraj E., Rafii-Tari H., Payne C. J., Azzie G., Sodergren M. H., et al. (2015). Force-sensing enhanced simulation environment (ForSense) for laparoscopic surgery training and assessment. Surgery 157 723–731. 10.1016/j.surg.2014.10.015 [DOI] [PubMed] [Google Scholar]
  17. Curtis M. T., Diazgranados D., Feldman M. (2012). Judicious use of simulation technology in continuing medical education. J. Contin. Educ. Health Prof. 32 255–260. 10.1002/chp.21153 [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. D’Angelo A.-L. D., Law K. E., Cohen E. R., Greenberg J. A., Kwan C., Greenberg C., et al. (2015). The use of error analysis to assess resident performance. Surgery 158 1408–1414. 10.1016/j.surg.2015.04.010 [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. De Blacam C., O’keeffe D. A., Nugent E., Doherty E., Traynor O. (2012). Are residents accurate in their assessments of their own surgical skills? Am. J. Surg. 204 724–731. 10.1016/j.amjsurg.2012.03.003 [DOI] [PubMed] [Google Scholar]
  20. Dogusoy-Taylan B., Cagiltay K. (2014). Cognitive analysis of experts’ and novices’ concept mapping processes: an eye tracking study. Comput. Hum. Behav. 36 82–93. 10.1016/j.chb.2014.03.036 [DOI] [Google Scholar]
  21. Doleck T., Jarrell A., Poitras E. G., Chaouachi M., Lajoie S. P. (2016). A tale of three cases: examining accuracy, efficiency, and process differences in diagnosing virtual patient cases. Australas. J. Educ. Technol. 36, 61–76. [Google Scholar]
  22. Earle D. (2006). Surgical training and simulation laboratory at Baystate Medical Center. Surg. Innov. 13 53–60. 10.1177/155335060601300109 [DOI] [PubMed] [Google Scholar]
  23. Eivazi S., Hafez A., Fuhl W., Afkari H., Kasneci E., Lehecka M., et al. (2017). Optimal eye movement strategies: a comparison of neurosurgeons gaze patterns when using a surgical microscope. Acta Neurochir. 159 959–966. 10.1007/s00701-017-3185-1 [DOI] [PubMed] [Google Scholar]
  24. Feldman M., Lazzara E. H., Vanderbilt A. A., Diazgranados D. (2012). Rater training to support high-stakes simulation-based assessments. J. Contin. Educ. Health Prof. 32 279–286. 10.1002/chp.21156 [DOI] [PMC free article] [PubMed] [Google Scholar]
  25. Fichtel E., Lau N., Park J., Parker S. H., Ponnala S., Fitzgibbons S., et al. (2019). Eye tracking in surgical education: gaze-based dynamic area of interest can discriminate adverse events and expertise. Surg. Endosc. 33 2249–2256. 10.1007/s00464-018-6513-5 [DOI] [PubMed] [Google Scholar]
  26. Gallagher A. G., Satava R. (2002). Virtual reality as a metric for the assessment of laparoscopic psychomotor skills. Surg. Endosc. 16 1746–1752. 10.1007/s00464-001-8215-6 [DOI] [PubMed] [Google Scholar]
  27. Gegenfurtner A., Lehtinen E., Jarodzka H., Säljö R. (2017). Effects of eye movement modeling examples on adaptive expertise in medical image diagnosis. Comput. Educ. 113 212–225. 10.1016/j.compedu.2017.06.001 [DOI] [Google Scholar]
  28. Gegenfurtner A., Lehtinen E., Säljö R. (2011). Expertise differences in the comprehension of visualizations: a meta-analysis of eye-tracking research in professional domains. Educ. Psychol. Rev. 23 523–552. 10.1007/s10648-011-9174-7 [DOI] [Google Scholar]
  29. Gegenfurtner A., Seppänen M. (2013). Transfer of expertise: an eye tracking and think aloud study using dynamic medical visualizations. Comput. Educ. 63 393–403. 10.1016/j.compedu.2012.12.021 [DOI] [Google Scholar]
  30. Grantcharov T. P., Bardram L., Funch-Jensen P., Rosenberg J. (2003). Learning curves and impact of previous operative experience on performance on a virtual reality simulator to test laparoscopic surgical skills. Am. J. Surg. 185 146–149. 10.1016/s0002-9610(02)01213-8 [DOI] [PubMed] [Google Scholar]
  31. Hoffmann E. R. (1997). Movement time of right-and left-handers using their preferred and non-preferred hands. Int. J. Ind. Ergon. 19 49–57. 10.1016/0169-8141(95)00092-5 [DOI] [Google Scholar]
  32. Horeman T., Dankelman J., Jansen F. W., Van Den Dobbelsteen J. J. (2013). Assessment of laparoscopic skills based on force and motion parameters. IEEE Trans. Biomed. Eng. 61 805–813. 10.1109/tbme.2013.2290052 [DOI] [PubMed] [Google Scholar]
  33. Horeman T., Rodrigues S. P., Jansen F. W., Dankelman J., Van Den Dobbelsteen J. J. (2011). Force parameters for skills assessment in laparoscopy. IEEE Trans. Haptics 5 312–322. 10.1109/toh.2011.60 [DOI] [PubMed] [Google Scholar]
  34. Hsu J. L., Korndorffer J. R., Jr., Brown K. M. (2015). Design of vessel ligation simulator for deliberate practice. J. Surg. Res. 197 231–235. 10.1016/j.jss.2015.02.068 [DOI] [PubMed] [Google Scholar]
  35. Iqbal S. T., Zheng X. S., Bailey B. P. (2004). “Task-evoked pupillary response to mental workload in human-computer interaction,” in Proceedings of the CHI’04 Extended Abstracts on Human Factors in Computing Systems, The Hague, 1477–1480. [Google Scholar]
  36. Jian Y.-C., Ko H.-W. (2017). Influences of text difficulty and reading ability on learning illustrated science texts for children: an eye movement study. Comput. Educ. 113 263–279. 10.1016/j.compedu.2017.06.002 [DOI] [Google Scholar]
  37. Jones D. B. (2007). Video trainers, simulation and virtual reality: a new paradigm for surgical training. Asian J. Surg. 30 6–12. 10.1016/s1015-9584(09)60121-4 [DOI] [PubMed] [Google Scholar]
  38. Kasarskis P., Stehwien J., Hickox J., Aretz A., Wickens C. (2001). “Comparison of expert and novice scan behaviors during VFR flight,” in Proceedings of the 11th International Symposium on Aviation Psychology, Columbus, OH, 1–6. [Google Scholar]
  39. Khan R. S., Tien G., Atkins M. S., Zheng B., Panton O. N., Meneghetti A. T. (2012). Analysis of eye gaze: do novice surgeons look at the same location as expert surgeons during a laparoscopic operation? Surg. Endosc. 26 3536–3540. 10.1007/s00464-012-2400-7 [DOI] [PubMed] [Google Scholar]
  40. Kruger J.-L., Doherty S. (2016). Measuring cognitive load in the presence of educational video: towards a multimodal methodology. Australas. J. Educ. Technol. 32 19–31. [Google Scholar]
  41. Lenhard W., Lenhard A. (2016). Calculation of Effect Sizes. Dettelbach: Psychometric. [Google Scholar]
  42. Loukas C., Nikiteas N., Kanakis M., Georgiou E. (2011). Deconstructing laparoscopic competence in a virtual reality simulation environment. Surgery 149 750–760. 10.1016/j.surg.2010.11.012 [DOI] [PubMed] [Google Scholar]
  43. MacDonald J., Williams R. G., Rogers D. A. (2003). Self-assessment in simulation-based surgical skills training. Am. J. Surg. 185 319–322. 10.1016/s0002-9610(02)01420-4 [DOI] [PubMed] [Google Scholar]
  44. Maran N. J., Glavin R. J. (2003). Low-to high-fidelity simulation–a continuum of medical education? Med. Educ. 37 22–28. 10.1046/j.1365-2923.37.s1.9.x [DOI] [PubMed] [Google Scholar]
  45. Mathis R., Watanabe Y., Ghaderi I., Nepomnayshy D. (2020). What are the skills that represent expert-level laparoscopic suturing? A Delphi Study. Surg. Endosc. 34 1318–1323. 10.1007/s00464-019-06904-w [DOI] [PubMed] [Google Scholar]
  46. McCrum-Gardner E. (2008). Which is the correct statistical test to use? Br. J. Oral Maxillofac. Surg. 46 38–41. 10.1016/j.bjoms.2007.09.002 [DOI] [PubMed] [Google Scholar]
  47. McDougall E. M., Corica F. A., Boker J. R., Sala L. G., Stoliar G., Borin J. F., et al. (2006). Construct validity testing of a laparoscopic surgical simulator. J. Am. Coll. Surg. 202 779–787. 10.1016/j.jamcollsurg.2006.01.004 [DOI] [PubMed] [Google Scholar]
  48. McLaughlin L., Bond R., Hughes C., Mcconnell J., Mcfadden S. (2017). Computing eye gaze metrics for the automatic assessment of radiographer performance during X-ray image interpretation. Int. J. Med. Inf. 105 11–21. 10.1016/j.ijmedinf.2017.03.001 [DOI] [PubMed] [Google Scholar]
  49. McNatt S., Smith C. (2001). A computer-based laparoscopic skills assessment device differentiates experienced from novice laparoscopic surgeons. Surg. Endosc. 15 1085–1089. 10.1007/s004640080022 [DOI] [PubMed] [Google Scholar]
  50. Meneghetti A. T., Pachev G., Zheng B., Panton O. N., Qayumi K. (2012). Objective assessment of laparoscopic skills: dual-task approach. Surg. Innov. 19 452–459. 10.1177/1553350611430673 [DOI] [PubMed] [Google Scholar]
  51. Menekse Dalveren G. G., Cagiltay N. E. (2019). Evaluation of ten open-source eye-movement classification algorithms in simulated surgical scenarios. IEEE Access 7 161794–161804. 10.1109/access.2019.2951506 [DOI] [Google Scholar]
  52. Mizota T., Anton N. E., Huffman E. M., Guzman M. J., Lane F., Choi J. N., et al. (2020). Development of a fundamentals of endoscopic surgery proficiency-based skills curriculum for general surgery residents. Surg. Endosc. 34 771–778. 10.1007/s00464-019-06827-6 [DOI] [PubMed] [Google Scholar]
  53. Moorthy K., Munz Y., Sarker S. K., Darzi A. (2003). Objective assessment of technical skills in surgery. BMJ 327 1032–1037. 10.1136/bmj.327.7422.1032 [DOI] [PMC free article] [PubMed] [Google Scholar]
  54. Munshi F., Lababidi H., Alyousef S. (2015). Low-versus high-fidelity simulations in teaching and assessing clinical skills. J. Taibah Univ. Med. Sci. 10 12–15. 10.1016/j.jtumed.2015.01.008 [DOI] [Google Scholar]
  55. Oostema J. A., Abdel M. P., Gould J. C. (2008). Time-efficient laparoscopic skills assessment using an augmented-reality simulator. Surg. Endosc. 22 2621–2624. 10.1007/s00464-008-9844-9 [DOI] [PubMed] [Google Scholar]
  56. Oropesa I., Sánchez-González P., Chmarra M. K., Lamata P., Fernández A., Sánchez-Margallo J. A., et al. (2013). EVA: laparoscopic instrument tracking based on endoscopic video analysis for psychomotor skills assessment. Surg. Endosc. 27 1029–1039. 10.1007/s00464-012-2513-z [DOI] [PubMed] [Google Scholar]
  57. Oropesa I., Sánchez-González P., Lamata P., Chmarra M. K., Pagador J. B., Sánchez-Margallo J. A., et al. (2011). Methods and tools for objective assessment of psychomotor skills in laparoscopic surgery. J. Surg. Res. 171 e81–e95. [DOI] [PubMed] [Google Scholar]
  58. Perrenot C., Perez M., Tran N., Jehl J.-P., Felblinger J., Bresler L., et al. (2012). The virtual reality simulator dV-Trainer® is a valid assessment tool for robotic surgical skills. Surg. Endosc. 26 2587–2593. 10.1007/s00464-012-2237-0 [DOI] [PubMed] [Google Scholar]
  59. Piccardi L., De Luca M., Nori R., Palermo L., Iachini F., Guariglia C. (2016). Navigational style influences eye movement pattern during exploration and learning of an environmental map. Front. Behav. Neurosci. 10:140. 10.3389/fnbeh.2016.00140 [DOI] [PMC free article] [PubMed] [Google Scholar]
  60. Reiley C. E., Lin H. C., Yuh D. D., Hager G. D. (2011). Review of methods for objective surgical skill evaluation. Surg. Endosc. 25 356–366. 10.1007/s00464-010-1190-z [DOI] [PubMed] [Google Scholar]
  61. Resnick R., Taylor B., Maudsley R. (1991). In-training evaluation—it’s more than just a form. Ann. R. Coll. Phys. Surg. Can. 24 415–420. [Google Scholar]
  62. Richstone L., Schwartz M. J., Seideman C., Cadeddu J., Marshall S., Kavoussi L. R. (2010). Eye metrics as an objective assessment of surgical skill. Ann. Surg. 252 177–182. 10.1097/sla.0b013e3181e464fb [DOI] [PubMed] [Google Scholar]
  63. Rosenthal R., Cooper H., Hedges L. (1994). Parametric measures of effect size. Handb. Res. Synth. 621 231–244. [Google Scholar]
  64. Rosser J. C., Jr., Rosser L. E., Savalgi R. S. (1998). Objective evaluation of a laparoscopic surgical skill program for residents and senior surgeons. Arch. Surg. 133 657–661. 10.1001/archsurg.133.6.657 [DOI] [PubMed] [Google Scholar]
  65. Schreuder H. W., Van Dongen K. W., Roeleveld S. J., Schijven M. P., Broeders I. A. (2009). Face and construct validity of virtual reality simulation of laparoscopic gynecologic surgery. Am. J. Obstet. Gynecol. 200 540.e1–540.e8. [DOI] [PubMed] [Google Scholar]
  66. Seagull F. J. (2012). Human factors tools for improving simulation activities in continuing medical education. J. Contin. Educ. Health Prof. 32 261–268. 10.1002/chp.21154 [DOI] [PubMed] [Google Scholar]
  67. Shetty S., Panait L., Baranoski J., Dudrick S. J., Bell R. L., Roberts K. E., et al. (2012). Construct and face validity of a virtual reality–based camera navigation curriculum. J. Surg. Res. 177 191–195. 10.1016/j.jss.2012.05.086 [DOI] [PubMed] [Google Scholar]
  68. Silvennoinen M., Mecklin J.-P., Saariluoma P., Antikainen T. (2009). Expertise and skill in minimally invasive surgery. Scand. J. Surg. 98 209–213. 10.1177/145749690909800403 [DOI] [PubMed] [Google Scholar]
  69. Stuijfzand B. G., Van Der Schaaf M. F., Kirschner F. C., Ravesloot C. J., Van Der Gijp A., Vincken K. L. (2016). Medical students’ cognitive load in volumetric image interpretation: insights from human-computer interaction and eye movements. Comput. Hum. Behav. 62 394–403. 10.1016/j.chb.2016.04.015 [DOI] [Google Scholar]
  70. Tien T., Pucher P. H., Sodergren M. H., Sriskandarajah K., Yang G.-Z., Darzi A. (2014). Eye tracking for skills assessment and training: a systematic review. J. Surg. Res. 191 169–178. 10.1016/j.jss.2014.04.032 [DOI] [PubMed] [Google Scholar]
  71. Tien T., Pucher P. H., Sodergren M. H., Sriskandarajah K., Yang G.-Z., Darzi A. (2015). Differences in gaze behaviour of expert and junior surgeons performing open inguinal hernia repair. Surg. Endosc. 29 405–413. 10.1007/s00464-014-3683-7 [DOI] [PubMed] [Google Scholar]
  72. Topalli D., Cagiltay N. E. (2019). Classification of intermediate and novice surgeons’ skill assessment through performance metrics. Surg. Innov. 26 621–629. 10.1177/1553350619853112 [DOI] [PubMed] [Google Scholar]
  73. Tricoche L., Ferrand-Verdejo J., Pélisson D., Meunier M. (2020). Peer presence effects on eye movements and attentional performance. Front. Behav. Neurosci. 13:280. 10.3389/fnbeh.2019.00280 [DOI] [PMC free article] [PubMed] [Google Scholar]
  74. Tsai M.-J., Huang L.-J., Hou H.-T., Hsu C.-Y., Chiou G.-L. (2016). Visual behavior, flow and achievement in game-based learning. Comput. Educ. 98 115–129. 10.1016/j.compedu.2016.03.011 [DOI] [Google Scholar]
  75. Uemura M., Jannin P., Yamashita M., Tomikawa M., Akahoshi T., Obata S., et al. (2016). Procedural surgical skill assessment in laparoscopic training environments. Int. J. Comput. Assist. Radiol. Surg. 11 543–552. 10.1007/s11548-015-1274-2 [DOI] [PubMed] [Google Scholar]
  76. Uhrich M., Underwood R., Standeven J., Soper N., Engsberg J. (2002). Assessment of fatigue, monitor placement, and surgical experience during simulated laparoscopic surgery. Surg. Endosc. 16 635–639. 10.1007/s00464-001-8151-5 [DOI] [PubMed] [Google Scholar]
  77. van der Lans R., Wedel M., Pieters R. (2011). Defining eye-fixation sequences across individuals and tasks: the Binocular-Individual Threshold (BIT) algorithm. Behav. Res. Methods 43 239–257. 10.3758/s13428-010-0031-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  78. van Empel P. J., Van Rijssen L. B., Commandeur J. P., Verdam M. G., Huirne J. A., Scheele F., et al. (2012). Validation of a new box trainer-related tracking device: the TrEndo. Surg. Endosc. 26 2346–2352. 10.1007/s00464-012-2187-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
  79. Van Sickle K. R., Ritter E. M., Smith C. D. (2006). The pretrained novice: using simulation-based training to improve learning in the operating room. Surg. Innov. 13 198–204. 10.1177/1553350606293370 [DOI] [PubMed] [Google Scholar]
  80. Vickers J. (1995). Gaze control in basketball foul shooting. Stud. Vis. Inf. Process. 6 527–541. 10.1016/s0926-907x(05)80044-3 [DOI] [Google Scholar]
  81. Wanzel K. R., Ward M., Reznick R. K. (2002). Teaching the surgical craft: from selection to certification. Curr. Probl. Surg. 39 583–659. 10.1067/mog.2002.123481 [DOI] [PubMed] [Google Scholar]
  82. Wilson M., Mcgrath J., Vine S., Brewer J., Defriend D., Masters R. (2010). Psychomotor control in a virtual laparoscopic surgery training environment: gaze control parameters differentiate novices from experts. Surg. Endosc. 24 2458–2464. 10.1007/s00464-010-0986-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  83. Yarbus A. L. (ed.) (1967). “Eye movements during perception of complex objects,” in Eye Movements and Vision, (Boston, MA: Springer; ), 171–211. 10.1007/978-1-4899-5379-7_8 [DOI] [Google Scholar]
  84. Zhang X., Jiang S., Ordóñez De Pablos P., Lytras M. D., Sun Y. (2017). How virtual reality affects perceived learning effectiveness: a task–technology fit perspective. Behav. Inf. Technol. 36 548–556. 10.1080/0144929x.2016.1268647 [DOI] [Google Scholar]
  85. Zheng B., Jiang X., Atkins M. S. (2015). Detection of changes in surgical difficulty evidence from pupil responses. Surg. Innov. 22 629–635. 10.1177/1553350615573582 [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The datasets generated for this study are available on request to the corresponding author.


Articles from Frontiers in Psychology are provided here courtesy of Frontiers Media SA

RESOURCES