Skip to main content
PLOS ONE logoLink to PLOS ONE
. 2020 May 29;15(5):e0223941. doi: 10.1371/journal.pone.0223941

Pupil diameter differentiates expertise in dental radiography visual search

Nora Castner 1,*, Tobias Appel 1,#, Thérése Eder 2,, Juliane Richter 2,, Katharina Scheiter 2,3,, Constanze Keutel 4,, Fabian Hüttig 5,, Andrew Duchowski 6,#, Enkelejda Kasneci 1,#
Editor: Susana Martinez-Conde7
PMCID: PMC7259659  PMID: 32469952

Abstract

Expert behavior is characterized by rapid information processing abilities, dependent on more structured schemata in long-term memory designated for their domain-specific tasks. From this understanding, expertise can effectively reduce cognitive load on a domain-specific task. However, certain tasks could still evoke different gradations of load even for an expert, e.g., when having to detect subtle anomalies in dental radiographs. Our aim was to measure pupil diameter response to anomalies of varying levels of difficulty in expert and student dentists’ visual examination of panoramic radiographs. We found that students’ pupil diameter dilated significantly from baseline compared to experts, but anomaly difficulty had no effect on pupillary response. In contrast, experts’ pupil diameter responded to varying levels of anomaly difficulty, where more difficult anomalies evoked greater pupil dilation from baseline. Experts thus showed proportional pupillary response indicative of increasing cognitive load with increasingly difficult anomalies, whereas students showed pupillary response indicative of higher cognitive load for all anomalies when compared to experts.

Introduction

Visual inspection is a commonly performed task in many contemporary professions, e.g. radiologists and other medical personnel frequently examine medical radiographs to diagnose and treat patients, airport security scan X-rays of luggage for prohibited items, etc. [1, 2]. In such tasks, expert visual inspection is derived from domain knowledge and is optimized for a short period of search. Thus, understanding the search process and measuring mental workload are fundamental in expert research towards developing computer-based metrics. Generally, visual performance, e.g. during search, has been characterized by metrics derived from the discrimination of fixations and saccades. Fixations are the period when eye movements are relatively still, indicating focus of attention, usually on areas prone to a specific goal [3, 4]. Saccades, the rapid eye movements, are usually made when scanning over irrelevant areas to a specific goal [5].

Of particular interest is estimation of cognitive load during visual search used in demanding real-world tasks. Images with complex features can affect performance, especially in visual search, and so selection of measurement techniques to assess human performance is paramount [6]. One especially important factor in performance is workload, where feature complexity has a measurable effect. This research focuses on the objective, non-invasive, physiological measure of cognitive load [7] via eye tracking. Consequently, we expect that cognitive load measures will manifest significant responses during the decision-making aspect of the visual search task.

We examined the differences between expert and novice inspectors of dental panoramic radiographs. Orthopantomograms (OPTs), which are information-dense 2D superimpositions of the maxillomandibular region and used frequently in all aspects of dental medicine [8]. Due to their heavy reliance on OPTs, dentists undergo professional training and licensing; however, they are still highly susceptible to under-detections and missed information [913]. Coupled with concern for patients’ health, accurate interpretation in spite of complex imagery is crucial. Specifically, OPTs have been shown to be less sensitive imagery for certain anomaly types than intraoral (periapical) radiographs, making correct detection more difficult [14, 15]. Therefore, less sensitive imagery of an anomaly can evoke higher gradation of difficulty for its accurate interpretation. Further understanding of both expert and novice OPT examination is necessary to effectively improve the training of medical image interpretation. Previous research has only scratched the surface of the cognitive processes during visual inspection of radiological images and the dichotomy between experts and novices. For this reason, our work goes one step further by examining the adaptability of cognitive processes during visual inspection of multiple features in decision making.

Background: Characterizing expertise

Expertise lies in the mind. The theory that expert aptitude develops a more structured long-term memory designated for domain-specific tasks [16] offers insight into experts’ faster and more accurate abilities [5]. Long-term working memory, proposed by Ericsson and Kintsch [16], offers this explanation for how experts seemingly effortlessly handle their domain-specific tasks. Their memory structuring facilitates their ability to maintain working memory at optimal capacity, avoiding overload, which affects productivity and performance.

Generally, working memory is understood as temporary storage for processing readily available information [17]. Long-term working memory relates to the structuring available to the larger, long-lasting storage and is of interest in skill learning [16]. For instance, chess players employ memory chunking that enables them to quickly recognize favorable positions and movements with less focus on single pieces [18]. Athletes show faster reaction to attentional cues, especially in interceptive sports, (e.g. basketball), indicating more rapid mental processing [19]. Also, medical professionals have been thought to proficiently employ heuristics in their decision-making strategies, i.e. visual search of radiographs [20] and diagnostic reasoning in case examinations [21, 22].

Developing new skills and the related memory structures for a specific discipline rely heavily on the capacity of working memory. According to Just and Carpenter [23], when the working memory demands exceed available capacity, comprehension is inhibited, leading to negative effects on performance. Effective comprehension then relies on resource allocation [23]. Optimal resource allocation supports rapid convergence to the most appropriate task-solution. Experts can filter out irrelevant information, which is evident in gaze behavior; they focus more on areas relevant to the task solution and less on areas that are irrelevant [5, 24, 25]. For instance, expert radiologists devote more fixations to anomaly-prone areas [26, 27] and devote shorter fixation time to an anomaly in detection tasks [20, 28]. Dental students’ gaze behavior has also been shown to be an effective feature to classify level of conceptual knowledge [29].

Additionally, when the task becomes too difficult or is perceived as such, there is more demand on working memory [30]. Sweller points out that the means-to-an-end problem solving strategies that novices employ can overload working memory [31]. And though perceived task-difficulty is influenced by acquired knowledge [32], even experts can face challenging problems that could evoke more load on working memory [33, 34]. Cognitive load, or more specifically intrinsic cognitive load [35], is the effect of “heavy use of limited cognitive-processing capability” [31]. For more information, see review by Paas and Ayres [36]. High cognitive load has been shown to have negative effects on performance [30] and effective learning in general [37].

One way to assess levels of cognitive load is the pupillary response [3840], where pupil size has been shown to increase as a response to memory capacity limits [41, 42] as well as when the task becomes too difficult [37, 43]. Accordingly, experts have a higher threshold for what is difficult compared to their novice counterparts, which is evident in the pupil response. Therefore, we are interested in expert and novice dentists when interpreting anomalies of varying degree of difficulty in panoramic radiographs. More important, our aim is to further understand experts’ perception of difficulty in their domain-specific tasks and whether this affects cognitive load.

Pupil diameter as a measure of cognitive load

Not only does visual search strategy reflect cognitive processes [4446], but pupil diameter has also been shown to be a robust, non-invasive measurement of cognitive load [3739, 4143, 4752]. Hence, with an increase in task difficulty, the diameter increases, otherwise known as task-evoked pupillary response. Originally, Kahneman and Beatty [47] linked pupil response to attentional differences. Then, the link between attention and capacity was promoted [43]; where higher load on the working memory showed a larger change in pupil dilation. Additionally, pupillary response has been found to be an indicator of learning [37], where pupil diameter decreased with more experience in a task.

Much of the early research in processing capacity and cognitive load has found that pupil activity correlates to workload during a variety of tasks [4143, 53]. Specifically for visual search tasks, cognitive load has also been measured by pupil activity. For instance, more distractors make the paradigm more difficult, affecting the pupil diameter increase [54]. Also, monochrome displays evoked longer search time and more pupil dilation than colored displays for both object counting and target finding tasks [55]. Regarding uncertainty, an increase in pupil diameter was associated with response time and uncertainty of target selection [56]. One of the more important takeaways from the visual search literature is the interplay of selective attention, increasing task demand, and the mental effort evoked. Moreover, this interplay is apparent in medical professionals and their diagnostic interpretation of radiographs. Students may not be as exposed to such tasks of varying difficulties, but accumulate more experiences overtime, which can reduce cognitive load. Regarding learning, pupil dilation decreases as an effect of training over time [57].

Though it is apparent that pupillary response is a product of cognitive load, other factors have been shown to effect pupil size, e.g. fatigue [58, 59], caffeine consumption [60], etc. [59, 61]. Most important to this work is changes in luminance in the environment, which result in the physiological response of constriction or dilation [52]. Age difference has also been shown to affect pupil size differences, where overall pupil size in older adults is smaller than younger adults, though variance between subjects in similar age groups is also quite high [48, 52]. With these factors in mind, studies on pupil diameter and load recommend a task-to-baseline comparison in luminance-controlled environments [3739, 4143, 47, 50, 54, 56, 62, 63]. Therefore, when measuring pupillary response in relation to cognitive load, these factors should be controlled in order to avoid such confounds.

Previous research

Only a few studies have comprehensively addressed cognitive load and medical expertise, and even fewer have addressed cognitive load during visual search. Trained physicians showed more accurate performance and smaller pupillary response during clinical multiple-choice questions compared to novices, and this effect was larger for more difficult questions [50]. Expert surgeons’ pupil diameter increased as a result of increasing task difficulty during laparoscopic procedures [64]. Additionally, Tien et al. [65] found that junior surgeons exhibited larger pupil sizes than experts during a surgical procedure. More important, they found that specific tasks affected junior surgeons’ pupillary response to a higher degree. For more references highlighting lower pupillary response as an effect of medical expertise (e.g. surgeons, anesthesiologists, physicians), see Szulewski et al. [66].

Regarding specifically medical image interpretation, Brunyé and colleagues [49] found pupil diameter increases as an effect of difficulty in diagnostic decision making, more so for cases that were accurately diagnosed. They further highlight the prospects that pupillary response in combination with gaze behavior has in understanding uncertainty in medical decision making [67]. Specifically for dental expertise and OPT interpretation, experts’ gaze behavior (e.g. fixations) was highly distinguishing of difficult and obvious images, where students’ gaze behavior was not [68, 69]. Castner et al. [13] found that fixation behavior changed with respect to differing anomalies. Therefore, the degree of difficulty in accurate pathology detection can affect gaze behavior, which can be indicative of the reasoning strategies used.

With this intention in mind, we looked at expert and novice dentists’ pupillary response while fixating on anomalies of varying difficulty in panoramic radiographs. To our knowledge, we are the first to apply differentiable pupillometry to the dental imagery visual search domain. Not only do these OPTs have multiple anomalies, but also within one OPT, varying difficulties can be present. Therefore, we are not analyzing an overall impression of easy or difficult image. Rather, through the course of the search strategy, we are extracting when dentists spot an anomaly and consequently mental processing at that moment. We propose the degree of anomaly interpretation difficulty can be indicated by changes in the pupillary response; where a larger response is more representative of harder to interpret anomalies. We also hypothesize to find a difference in the pupillary response between experts and novices, as established by prior research; where baseline-related pupil difference, as a measure of cognitive load, is sensitive to experts’ processing of anomalies of varying degree of difficulty. Additionally, we report that students, after acquiring the appropriate training to inspect OPTs, have higher cognitive load compared to experts. More interesting is whether students are attuned to the varying gradations of the anomalies.

Materials and methods

Participants

Data collection took place in the context of a larger project performed over multiple semesters from 2017 to 2019. Dentistry students from semesters six through ten were recorded during an OPT inspection task. For reference, sixth semester students are in the second half of their third year and the tenth semester is in the fifth year of their studies, being the last semester before they continue on to the equivalent of a residency.

The sixth semester students were evaluated three times in each period of data collection due to their curriculum requirement of an OPT interpretation training course. For the purpose of the present paper, we chose to only evaluate the sixth semester students after this course (Nsixth = 50). They have the necessary knowledge to perform the OPT task as it is intended (i.e. they know what they have to look for), without having yet acquired the routine skills.

Table 1 details both the student and expert data. Experts (Nexperts = 28) from the University clinic volunteered their expertise for the same task that students performed. Experience was defined as professional years working as a dentist and ranged from 1 to 43 years (Myears = 9.88). 50% of experts reported seeing between 11 and 30 patients on a typical work day and the remainder saw less than 10 patients a day. All experts had the necessary qualifications to practice dentistry and or any other dental related specialty: e.g. Prosthodontics, Orthodontics, Endodontics, etc. Due to technical difficulties, eye tracking data was lost for two participants, leaving Nexperts = 26 participants for the eye tracking analysis.

Table 1. Participant data overview.

Students Experts
N 50 26
Nglasses 12* 9
OPTs viewed/person 20 15
Total Datasets 750 390
Poor Tracking Ratio** 14.3% 14.3%

* data regarding glasses for one collection is unknown

** Percentage of poor data quality. Proportion of valid gaze points less than 80%.

The Ethical Review Board of the Leibniz-Institut für Wissensmedien Tübingen approved the student cohort of the study with the project number LEK 2017/016. All participants were informed in written form and consented in written form that their pseudonymous data can be analyzed and published. Due to a self-constructed pseudonym, they had the option to revoke this consent until the date of anonymization of the data after data collection is finished. The Independent Ethics Committee of the Medical Faculty and University Hospital Tübingen approved the expert cohort of the study with the project number 394/2017BO2. All participants were informed in written form and consented verbally that their anonymous data can be analyzed and published. Due to a self-constructed pseudonym, they had the option to revoke this consent at any time.

Experimental paradigm

The experimental protocol for the students consisted of an initial calibration, task instruction, then two image phases: Interpretation and Marking. The details of the experimental protocol are found in Fig 1. Prior to the interpretation, a two second fixation cross was presented: This served as baseline for our analysis. Then, an OPT was presented in the interpretation phase for 90 seconds and the participant was instructed to only search for areas indicative of any pathologies in need of further intervention. The marking phase came after interpretation; where the same OPT was shown with the instruction to only mark the anomalies found in the interpretation phase using an on-screen drawing tool. There was unlimited time for the marking phase and participants could continue with a button click. This procedure was repeated for all OPTs. In total, the students viewed 20 OPTs with a short break after the first ten.

Fig 1. Outline of experimental session.

Fig 1

Initially, there was a calibration and procedural instructions. Then for each image, there is a fixation cross for baseline data, the exploration phase (45s duration for experts and 90s for students), instructions for the marking phase, and the marking phase (unlimited time). Students received two sets of 10 OPTs with a break in between and experts received one set of 15 OPTs with a break after the first seven.

The diagnostic task for the expert group was highly similar to that of the students. However, it was determined that 90 seconds is too long of a duration for the experts, since much of the previous literature has shown experts are faster at scanning radiographs [5, 20, 26, 27, 68, 7072]. Therefore, the exploration phase was shortened to a duration of 45 seconds. Additionally, due their busy schedules, experts only viewed 15 OPTs, with a short pause after the first seven.

Both students and experts were unrestrained during the experiment, although they were instructed to move their head as little as possible. Further details of one of the student data collections can be found in Castner et al. [29] and expert data collections can be found in Castner et al. [13].

Stimuli

OPT images

The 15 OPTs viewed by both the experts and the post-training course sixth semester students were used for the current analysis to avoid effects from unseen images. The OPTs were chosen from the university clinic database by the two expert dentists involved in this research project and were determined to have no artifacts and technological errors. Both dentists independently examined the OPTs and the patient workups and further consolidated together to determine ground truths for each image. Two OPTS were negative (no anomalies) controls.

Additionally, the level of difficulty for each anomaly was pre-determined. Fig 2 shows three OPT images viewed in the experiment. Anomalies are illustrated in green, yellow, and red, and represent easy, medium, and difficult, respectively. This classification was set up in a blinded review and the consent process of two senior dentists (6th and 7th authors). For example, the green anomalies in Fig 2A are dental cyst (1) and insufficient root canal fillings. (2a,b) in Fig 2C are an example of elongated lower molars due to missing antagonists. The yellow anomalies in Fig 2B are irregular forms of the mandibular condyle (1,3) and (2) is an apical translucency indicative of inflammation due to a contagious (bacterially colonized) root canal filling. The red anomalies in this image are approximal caries (4) and a maxillary sinus mass. Anomalies indicated by the white dashed circles were determined as ambiguous, e.g. the nature of their difficulty and or pathology is unclear. For example, in Fig 2B (7,8) are impacted wisdom teeth, though it is uncertain whether this will become a problem for the patient and therefore is regarded as potentially pathologic. (6) is an apical translucency at the mesial root apex and it is unclear whether it is indicative of an inflammation. Therefore, they were kept in this analysis even though the nature of their difficulty is unclear.

Fig 2. OPTs with pre-determined ground truth.

Fig 2

Example of the OPTs used in the experiment. Pre-determined ground truths are indicated by the ellipses and their colors indicate the level of difficulty each anomaly is: Green (least difficult), yellow (intermediary), red (most difficult) and white (nature of difficulty unclear). Image (D) is the ground truth map for image (B). Each anomaly is segmented and given a distinguishing integer.

Ground truth maps

We created maps for the 15 OPTs evaluated (See Fig 2C) using Matlab 2018. As input, all OPTs were loaded as .png files with their respective anomalies—all colored red. Thresholding for red values was performed to automatically get the pixel coordinates of the ellipse edges. Then, the ellipses were filled with the poly2mask() function. Anomalies automatically extracted from this process were double checked for overlapping and had their boundaries corrected. Similar anomalies inside of another, such as (2a,b) in Fig 2C, were grouped together as one anomaly. Other anomalies too close together and too different in pathology, such as (3,8) in Fig 2C, were excluded from the analysis, due to possible spatial accuracy errors in the gaze. Similarly, anomalies that were denoted by too small of an ellipse were padded to have a larger pixel area, e.g. (4) in Fig 2B, to account for the spatial accuracy errors in the gaze. Each segmented anomaly is given a distinguishing integer for its respective pixels. Raw gaze points from the left eye are then mapped to the map and gaze coordinates receive the corresponding integer value.

Data acquisition

Environment

Data collection for students took place in a digital classroom equipped with 30 remote eye trackers attached to laptops with 17inch HD display screens running at full brightness. This setup allows for data collection of up to 30 participants simultaneously, minimizing the overall time needed for collection. For this study, verbal instructions were given en masse pertaining to a brief overview of the protocol and an explanation of eye tracking, then individual calibrations were performed with a supervised quality check; students could then run the experiment self-paced.

Data collection for the experts took place in the university hospital so the experts could conveniently participate during work hours. There, the room used for data collection was dedicated for radiograph reading. The same model remote eye tracker was used for expert data collection and was run with the same sampling frequency on a laptop with 17inch HD display screen running at full brightness.

More important to the current study, both data collection environments had the room illumination levels controlled with no effects from sunlight or other outdoor light. The standard maintained illuminance for experimental sessions was between 10 to 50 lux, measured with a lux sensor (Gossen Mavo-Max illuminance sensor, MC Technologies, Hannover, Germany). It is advised that environment illumination during radiograph reading should be ambient (25–50 lux) for the best viewing practices [73] and to optimize contrast perception in radiographs [7476]. Therefore, with room illumination controlled, we can evaluate pupillary response independent of environmental illumination changes.

Laptops

Regarding the screen display, radiograph reading is not affected by the luminance of the display [75]. However, both the laptop models used for the experimental sessions abided by the multiple medical and radiology commission standards [72, 73, 77]. The HP Z Book 15 (for students) has screen brightness averages approx. 300cd/m2 [78]. The Dell Precision m4800 (for experts) averages approx. 380cd/m2 [79]. While the screen luminance was also controlled and followed the standard protocols for viewing radiographs, the exact effect of the screen brightness on the pupillary response is out of the scope of this work; rather the pupillary response dependent on mental load during these reading task is the focus.

Eye tracker

The SMI RED250 remote eye tracker is a commercial eye tracker with 250Hz sampling frequency and used for gaze data collection. We used the included software for both the experiment design (Experiment Center) and event analysis (BeGaze). Since the eye tracker has a high sampling frequency, both stable (fixations) and rapid (saccadic) eye movements for static stimuli can be measured. Analysis was performed on the raw gaze data output from the eye tracker: x and y coordinates with timestamps mapped to the screen dimensions. The raw data points also have pupil diameter output in millimeters [80]. Although the data is raw and has not been run through event detection algorithms, raw gaze points are labeled as fixation, saccade, or blink.

Calibration was performed for all participants. A validation also was performed as a quality check to measure the gaze deviation for both eyes from a calibration point: A deviation larger than one degree constituted recalibration. Calibrations were performed prior to the experiments as well as one or two times during the experimental session, depending on how many images were presented.

Data preprocessing

Quality of raw data

Only gaze data from the exploration phase was of interest to this work since gaze data from the marking phase was affected by the use of the screen drawing-tool. Initially, the raw gaze data was examined for signal quality. The eye tracker reports proportion of valid gaze signal to stimulus time as the tracking ratio. Therefore, if a participant’s tracking ratio for an OPT was deemed insufficient—less than 80%—we omitted his or her data for this OPT. If overall, a participant had poor tracking ratios for more than three of OPTs he or she viewed, all gaze data for that participant was removed. This preprocessing stage can assure that errors (e.g. post-calibration shifts, poor signal due to glasses) in the gaze data are substantially minimized. Table 1 gives the distribution of participants and the percent of datasets excluded due to low tracking ratio (last row). We started with 1140 data sets, but 199 datasets were initially excluded on the grounds of poor data quality.

Blink removal

The SMI-reported tracking ratio does not take into account when the eye tracker detects a blink [80]. Nevertheless, inaccurately detected blinks created an alarming number of cases with acceptable tracking ratios even though there was an inordinate amount of undetected gaze. Fig 3a shows an example of a participant’s pupil size samples over time for the left and right eye for an OPT presentation. This participant had a reported tracking ratio of 98%, but a large portion of the left eye gaze signal– approximately 33.5 seconds out of 90 seconds—could be signal loss labeled as a blink. In contrast, Fig 3b shows a participant who also has a high tracking ratio, though the data appears to be acceptable with typical blink durations detected and little signal loss.

Fig 3. Blink detection in the raw gaze data.

Fig 3

(a) Low Data Quality Example (b) High Data Quality Example. The raw pupil signal of the left and right eye (orange and blue dots) over the course of image presentation. Red and green dots in the lower part show when the eye tracker labels the data point as a blink for the left and right eye, respectively. The particular subject in 3a had a high tracking ratio, though many data samples could be incorrectly labeled as blinks. The participant in 3b also has a high tracking ratio and his or her data appears to be acceptable quality.

Consequently, the main issue stems from the apparent lack of a maximum blink duration threshold. Extra criteria were necessary to further detect and exclude datasets with pupil signal loss mislabeled as a blink. We overestimated the threshold for atypical blink durations, setting this value to 5000 ms, to account for situations where a participant could possibly be rubbing his or her eye/s or even closing the eye shortly. This threshold optimally maintains an acceptable amount of pupil data for the entire stimulus presentation (90 or 45 seconds). Since baseline data was sampled during the two seconds the fixation cross was displayed, we set the threshold blink duration to 500 ms and added an extra criterion of a minimum 200 pupil samples to effectively extract enough samples for an acceptable pupil diameter baseline. Therefore, 570 datasets from 72 participants (48 students, 24 experts) were used for the final analysis.

Pupil diameter measurement

Data analysis was done for the left eye. For further signal processing, we removed gaze coordinates and pupil data for the raw data points labeled as saccades (since visual input is not perceived during rapid eye movements [3]). Data points with a pupil diameter of zero or labeled as a blink were also removed. Additionally, data points 100 ms before and after blinks were removed, due to pupil size distortions from partial eye-lid occlusion. Lastly, the first and last 125 data points in the stimulus presentation were removed due to stimulus flickering [8183]. The remaining data was smoothed with a third order low-pass Butterworth filter with a 2Hz cutoff as illustrated by the purple data points in Fig 4.

Fig 4. Smoothed pupil signal.

Fig 4

Raw signal from the left eye (orange) and the smoothed signal (purple) with a Butterworth filter with 2Hz cutoff.

Gaze hit mapping

For both students and experts, we plotted the raw gaze points that landed in each anomaly and extracted its level of difficulty. For simplicity, we will refer to them as gaze hits. For all hits on an anomaly for a participant, we calculated the median pupil diameter. The median pupil diameter for each anomaly was then subtracted from the respective baseline data for that image. We performed subtractive baseline correction because it has been found to be a more robust metric and have higher statistical power [63]. Therefore, the difference from baseline could indicate diameter increase (positive value) or diameter decrease (negative value) compared to baseline.

With the gaze hits on anomalies of varying difficulties, we can evaluate the pupillary response of both experts and students during anomaly fixations. The pupillary response, as measured by change from baseline, can then provide insight into the mental/cognitive load both groups are undergoing while interpreting the anomalies.

Results

Overall change from baseline

Independent of gaze on anomaly difficulty, we looked at participants’ median pupil diameter for each image compared to baseline median pupil diameters. We favored the median over the mean because it has greater robustness towards noise and outliers. Fig 5a shows the average of the median pupillary response from baseline for both students and experts. Overall, students (M = 0.314, SD = 0.315) had a larger increase from baseline than experts (M = 0.057, SD = 0.353: t(568) = −8.824, p < 0.001). We also performed a supplementary analysis to rule out any effects that fatigue could have on the pupillary response (see S1 Fig).

Fig 5. Pupillary response of experts and novices during visual Inspection.

Fig 5

(a) Median Pupil Change From Baseline for Experts and Novices. (b) Median Pupil Change From Baseline for Gaze on Anomalies. The median pupil diameter change from baseline for students (blue bars) and experts (red bars) for the overall image behavior (5a) and when gazing on anomalies of varying difficulty (5b). Standard errors are indicated in black. Students had larger pupillary response from baseline compared to experts, but this effect was homogeneous for the differing anomalies. Whereas experts showed an increased pupillary response behavior as an effect of increasing difficulty.

Gaze on anomalies

To evaluate whether anomaly difficulty had an effect of student and expert pupillary response, we ran a 2 × 4 factor ANOVA to test for expertise and anomaly difficulty interactions. There was a main effect for expertise (F(1, 1388) = 161.68, p < 0.001) indicating that students had a larger increase from baseline than experts. There was also an effect for anomaly difficulty (F(3, 1388) = 3.87, p = 0.009) indicating that there was a larger increase in pupil size from baseline for more difficult anomalies. There was a significant interaction between expertise and anomaly difficulty (F(3, 1388) = 2.76, p = 0.041). There were no significant effects of anomaly difficulty on student pupillary response. However, there were significant effects of anomaly difficulty on expert pupillary response. Fig 5b details the pupillary response of experts and novices on the varying anomaly difficulties.

Post hoc analyses with Bonferroni correction for anomaly difficulty on the expert data revealed significant differences for the more difficult anomalies (M = 0.246, SD = 0.370) compared to least difficult (M = 0.0514, SD = 0.396, t(207) = −3.0582, p = 0.003) and ambiguous (t(150) = 3.1796, p = 0.002). There were no significant differences for medium anomalies (M = 0.1259, SD = 0.3904) compared to the difficult (t(200) = 1.8989, p = 0.059). Meaning, experts had the largest pupil size change from baseline for more difficult anomalies, especially compared to least difficult and ambiguous anomalies.

Discussion

Students showed larger and more homogenous pupil size change from baseline for all anomaly gradations compared to experts. Thus for students, pupillary response was independent of whether an anomaly was easy or difficult to interpret. This effect was also found during visual inspection of the whole image (Fig 5a), where students had overall greater change from baseline compared to experts. Pupillary response differences between students and experts have been supported by the previous literature [49, 50, 6567, 84]. However, the more interesting takeaway from this work is the lack of influence of anomaly gradation on student cognitive processing. One would imagine that even the most pronounced of anomalies would make the recognition process easier. Our findings from student pupillary response indicate that, regardless of how conspicuous, the level of mental workload remains constant.

Conversely, experts showed a strong pupillary response to anomaly gradation. The least difficult to interpret anomalies showed less change from baseline, then the intermediary anomalies, and finally the largest response was for the most difficult anomalies (Fig 5b). Meaning, as the gradation of difficulty increases so does the pupillary response. This behavior, however, was not evident for the ambiguous anomalies, which showed the smallest response change from baseline. This effect may lie in the nature of the uncertainty of these anomalies. As determined by the two experts involved in the project, this category was a mixture of potential areas that may or may not have included an anomaly: Or even an anomaly, but with no cause for alarm. Therefore, it is uncertain how difficult, easy, or even existing these anomalies were.

Cognitive load is often used to explain findings regarding learning [23, 31, 36, 62]. For instance, Tien et al. [65] found that novices reported higher memory load compared to experts performing the same task. This behavior can be likened to students’ lack of conceptual knowledge and experience, producing them to “think harder” [85, 86] to interpret these images. Furthermore, large pupil size can be reflective of learning during the task [23, 37, 41, 43, 47, 82]. During learning, students are developing the proper memory structures as theorized by Ericsson and Kintsch [16] and Sweller [31]. Additionally, their pupillary response could reflect that they have not yet developed the conceptual knowledge to quickly recognize the image features indicative of the specific anomalies or how to interpret their underlying pathologies. Even for easy anomalies, they may be unsure of whether they accurately interpreted it or not. Uncertainty as well as perceived task difficulty have been found to affect the pupillary response, and acquired knowledge has been shown to reduce uncertainty and perceived difficulty [32, 56]. Moreover, prior knowledge to a problem has been shown to reduce cognitive load [31, 36, 41, 50].

Cognitive load can also be indicative of inefficient reasoning strategies. Efficient reasoning strategies reduce load on working memory, in turn enhancing performance [30]. Patel et al. [33] found that when novices interpreted clinical case examinations, they tended to employ reasoning strategies that have been known to elicit higher workload. Our findings also suggest that students may employ similar cognitive strategies that evoke higher load for all anomaly gradations. Comparatively, experts employ more efficient strategies; however, they are more sensitive to task features.

In general, as task difficulty increases, so does the workload [64] and correspondingly, the pupil dilation [30, 43, 87, 88]. With increasingly difficult stimuli, Duchowski et al. [89] also showed increased cognitive load via microsaccade rates during decision making. However, Patel et al. [34] found more cognitive load in physicians when examining more complicated case examinations. When expert dentists perform a visual inspection of an OPT, they gaze in many areas that potentially have a multitude of differing pathologies or even positional and summation errors. Depending on the gradation of the area they are focusing on, proper interpretation may need to evoke adaptations in the decision-making strategies. Our findings show that experts dentists are capable of this adaptability during the course of visual inspection of OPTs.

Gaze behavior in expert dentists was also shown to change with difficult images [13, 68]. The current work went one step further and found changes within the visual search of an OPT in contrast to the overall response to image interpretation. Kok et al. [46] found that expertise reflected visual search strategies employed. Top-down strategies that experts generally employ use acquired knowledge and understanding of the current problem to focus on the relevant aspects of an image to quickly and more accurately process it [24, 90, 91]. Whereas bottom-up strategies that student generally employ is less efficient, as focus is on salient, noticeable images features, regardless of relevancy [20, 46, 91]. Furthermore, systematic search (inspecting all features of an image in a pre-determined orders) evokes more load on the working memory [20, 27]. However, students are generally trained to perform this type of search when they first get exposed to these images [72, 90].

An expert generally knows in what areas of the OPT anomalies are prevalent and how they are illustrated in the image features. Therefore, an expert can quickly recognize an image feature as a specific anomaly. In contrast to overall visual inspection—where experts showed low pupillary response compared to students—when inspecting specific areas, pupil dilation fluctuation can be indicative to changes in their cognitive processes to accommodate more complex features. Naturally, interpretation of medical images is not trivial and certain image or pathology features can avert the true diagnosis. Experts are more robust at determining more difficult or subtle anomalies [11, 27, 68, 72, 92]. Although when anomalies become harder to interpret, experts evoke pupillary response indicative of increasing task-difficulty, leading to behavior that is likely of more thorough inspection.

Limitations and future work

It should be noted that there were age differences between the two groups. Due to the sensitivity of the expert demographic data, we did not record their ages; but we expect them to be older than their student counterparts. Age has been found to have an effect on the average pupil size [48, 52]. For this reason, we measured a change from baseline to control such for age effects. Additionally, Van Gerven et al. [51] found that pupillary response to workload in older adults (early seventies) is not as pronounced as in younger adults (early twenties). Though we cannot say exactly how old our expert population was, they were all still working in the clinic and therefore more than likely to be younger than early seventies. Also, their years of experience in the clinic (average of 10 years) suggests they were more middle aged (30 to 45 years old). Further research is needed to better address this limitation and control for possible age difference effects on pupillary response.

Another limitation to this work could be the technical problems associated with the eye tracker data collection. We systematically removed data sets determined as poor quality; however, spatial resolution errors can accumulate within an experimental session if a participant moves too much. Then, the gaze appears to have a shifted offset, which would affect precision. Multiple calibrations during collection help with precision. We also increased the areas of smaller ground-truth anomalies and excluded anomalies that were too close and too different in nature. The total gaze hits on each type of anomaly were not evenly distributed, with more gaze hits on easier and intermediary anomalies (See S1 Table in Supporting Information). Students used more total gaze hits due to longer OPT presentation time, but the distributions were highly similar to experts. Future research could further untangle the differences in gaze hits on easier and difficult anomalies, while controlling for presentation time differences.

The temporal scanpath information is also an interesting direction for future research, i.e. systematic search in students and its effect on workload and pupillary response. For example, how often do “look backs” on anomaly areas occur and does the pupil dilation increase with each look back. Also, whether easy or more conspicuous anomalies are viewed at first and how the pupillary response in students incorporates this initial information. Following up on the understanding that systematic search produces more memory load as measured by pupil dilation [93], would also be interesting to replicate with temporal information from our findings.

Conclusion

We measured pupil diameter change from baseline when gazing on anomalies of varying difficulty during visual search of dental panoramic radiographs. We found that the gradation of anomalies in these images had an effect on expert pupillary response. Anomaly gradation did not have an effect on student pupillary response, which suggests higher workload and less sensitivity to complex features compared to experts. Experts are able to selectively allocate their attention to relevant information and is evident in the pupillary response. However, selective attention coupled with focus on features perceived as challenging can increase the pupil dilation as we found in our investigation. Although a majority of expert studies have established that experts are more robust at accurately solving their domain-specific tasks than their student counterparts [5, 16, 24, 91], increased pupillary response during difficult anomaly inspection supports adaptable processing strategies.

With more insight into expert decision-making processes during visual search or medical images, appropriate learning interventions can be developed. These interventions can incorporate not only the scanpath behavior, but also the cognitive load during appropriate detection of pathologies. From this combination, image semantics can be better conveyed to the learner. Training sessions that convey the appropriate information through adaptive gaze interventions based on cognitive load detection via the pupillary response offers a promising direction in medical education.

Supporting information

S1 Fig. Pupillary response over course of experiment.

The average pupillary response from baseline for students (blue bars, 20 images total) and experts (red bars, 15 images total) during the first set of OPTs presented and the second set of OPTs presented. Their is no effect in the pupillary response that could be attributed to fatigue during the experiment.

(PDF)

S1 Table. Table of Expert and Student Gaze Counts.

Shows the gaze hits on each anomaly type for both students and experts. For both levels of expertise, the least difficult and intermediate have the most gaze hits. The following are the ambiguous and the most difficult anomalies. Students had overall more gaze hits than experts; however, this may be attributed to the 90 second viewing time they had in comparison to the 45 second viewing time that the experts had.

(PDF)

Data Availability

The data and analysis scripts from the presented study are publicly available at: ftp://peg-public:peg-public@messor.informatik.uni-tuebingen.de/peg-public/norac/VisualExpertiseRadiology.zip.

Funding Statement

The student study is funded by the WissenschaftsCampus "Cognitive Interfaces" Tübingen (Principle Investigators: KS, CK and FH). The expert study with specialists and part of the data evaluation runs on budget of the University Hospital Tübingen / Department of Prosthodontics (Eberhard Karls University). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Qvarfordt P, Biehl JT, Golovchinsky G, Dunningan T. Understanding the Benefits of Gaze Enhanced Visual Search. In: Proceedings of the 2010 Symposium on Eye-Tracking Research & Applications. ETRA’10. New York, NY: ACM; 2010. p. 283–290. Available from: http://doi.acm.org/10.1145/1743666.1743733.
  • 2.Otto K, Castner N, Geisler D, Kasneci E. Development and evaluation of a gaze feedback system integrated into eyetrace. In: Proceedings of the 2018 ACM Symposium on Eye Tracking Research & Applications; 2018. p. 1–5.
  • 3. Holmqvist K, Nyström M, Andersson R, Dewhurst R, Jarodzka H, Van de Weijer J. Eye tracking: A comprehensive guide to methods and measures. OUP Oxford; 2011. [Google Scholar]
  • 4.Tafaj E, Kasneci G, Rosenstiel W, Bogdan M. Bayesian online clustering of eye movement data. In: Proceedings of the symposium on eye tracking research and applications; 2012. p. 285–288.
  • 5. Gegenfurtner A, Lehtinen E, Säljö R. Expertise differences in the comprehension of visualizations: A meta-analysis of eye-tracking research in professional domains. Educational Psychology Review. 2011;23(4):523–552. 10.1007/s10648-011-9174-7 [DOI] [Google Scholar]
  • 6. Moacdieh N, Sarter N. Display Clutter: A Review of Definitions and Measurement Techniques. Human Factors. 2015;57(1):61–100. 10.1177/0018720814541145 [DOI] [PubMed] [Google Scholar]
  • 7. Gwizdka J. Distribution of Cognitive Load in Web Search. J Am Soc Inf Sci Technol. 2010;61(11):2167–2187. 10.1002/asi.21385 [DOI] [Google Scholar]
  • 8. Huettig F, Axmann D. Reporting of dental status from full-arch radiographs: Descriptive analysis and methodological aspects. World Journal of Clinical Cases: WJCC. 2014;2(10):552 10.12998/wjcc.v2.i10.552 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9. Baghdady MT, Pharoah MJ, Regehr G, Lam EW, Woods NN. The role of basic sciences in diagnostic oral radiology. Journal of dental education. 2009;73(10):1187–1193. [PubMed] [Google Scholar]
  • 10. Baghdady MT, Carnahan H, Lam EW, Woods NN. Dental and dental hygiene students’ diagnostic accuracy in oral radiology: effect of diagnostic strategy and instructional method. Journal of dental education. 2014;78(9):1279–1285. [PubMed] [Google Scholar]
  • 11. Diniz MB, Rodrigues JA, Neuhaus K, Cordeiro RCL, Lussi A. Influence of Examiners’ Clinical Experience on the Reproducibility and Validity of Radiographic Examination in Detecting Occlusal Caries. Clinical Oral Investigations. 2010;14(5):515–523. 10.1007/s00784-009-0323-z [DOI] [PubMed] [Google Scholar]
  • 12. Bruno MA, Walker EA, Abujudeh HH. Understanding and confronting our mistakes: the epidemiology of error in radiology and strategies for error reduction. Radiographics. 2015;35(6):1668–1676. 10.1148/rg.2015150023 [DOI] [PubMed] [Google Scholar]
  • 13.Castner N, Klepper S, Kopnarski L, Hüttig F, Keutel C, Scheiter K, et al. Overlooking: the nature of gaze behavior and anomaly detection in expert dentists. In: Proceedings of the Workshop on Modeling Cognitive Processes from Multimodal Data. ACM; 2018. p. 8.
  • 14. Douglass CW, Valachovic RW, Wijesinha A, Chauncey HH, Kapur KK, McNeil BJ. Clinical efficacy of dental radiography in the detection of dental caries and periodontal diseases. Oral Surgery, Oral Medicine, Oral Pathology. 1986;62(3):330–339. 10.1016/0030-4220(86)90017-4 [DOI] [PubMed] [Google Scholar]
  • 15. Akarslan Z, Akdevelioglu M, Gungor K, Erten H. A comparison of the diagnostic accuracy of bitewing, periapical, unfiltered and filtered digital panoramic images for approximal caries detection in posterior teeth. Dentomaxillofacial Radiology. 2008;37(8):458–463. 10.1259/dmfr/84698143 [DOI] [PubMed] [Google Scholar]
  • 16. Ericsson KA, Kintsch W. Long-term working memory. Psychological review. 1995;102(2):211 10.1037/0033-295X.102.2.211 [DOI] [PubMed] [Google Scholar]
  • 17. Cowan N. What are the differences between long-term, short-term, and working memory? Progress in brain research. 2008;169:323–338. 10.1016/S0079-6123(07)00020-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18. Charness N, Reingold EM, Pomplun M, Stampe DM. The perceptual aspect of skilled performance in chess: Evidence from eye movements. Memory & cognition. 2001;29(8):1146–1152. 10.3758/BF03206384 [DOI] [PubMed] [Google Scholar]
  • 19. Voss MW, Kramer AF, Basak C, Prakash RS, Roberts B. Are expert athletes ‘expert’in the cognitive laboratory? A meta-analytic review of cognition and sport expertise. Applied Cognitive Psychology. 2010;24(6):812–826. 10.1002/acp.1588 [DOI] [Google Scholar]
  • 20. Van der Gijp A, Ravesloot C, Jarodzka H, van der Schaaf M, van der Schaaf I, van Schaik JP, et al. How visual search relates to visual diagnostic performance: a narrative systematic review of eye-tracking research in radiology. Advances in Health Sciences Education. 2017;22(3):765–787. 10.1007/s10459-016-9698-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21. Coderre S, Mandin H, Harasym PH, Fick GH. Diagnostic reasoning strategies and diagnostic success. Medical education. 2003;37(8):695–703. 10.1046/j.1365-2923.2003.01577.x [DOI] [PubMed] [Google Scholar]
  • 22. Schmidt HG, Boshuizen HP. On acquiring expertise in medicine. Educational psychology review. 1993;5(3):205–221. 10.1007/BF01323044 [DOI] [Google Scholar]
  • 23. Just MA, Carpenter PA. A capacity theory of comprehension: individual differences in working memory. Psychological review. 1992;99(1):122 10.1037/0033-295X.99.1.122 [DOI] [PubMed] [Google Scholar]
  • 24. Haider H, Frensch PA. Eye movement during skill acquisition: More evidence for the information-reduction hypothesis. Journal of Experimental Psychology: Learning, Memory, and Cognition. 1999;25(1):172. [Google Scholar]
  • 25. Eivazi S, Hafez A, Fuhl W, Afkari H, Kasneci E, Lehecka M, et al. Optimal eye movement strategies: a comparison of neurosurgeons gaze patterns when using a surgical microscope. Acta Neurochirurgica. 2017;159(6):959–966. 10.1007/s00701-017-3185-1 [DOI] [PubMed] [Google Scholar]
  • 26. Nodine CF, Mello-Thoms C. The nature of expertise in radiology Handbook of Medical Imaging SPIE; 2000. [Google Scholar]
  • 27. Kok EM, Jarodzka H, de Bruin AB, BinAmir HA, Robben SG, van Merriënboer JJ. Systematic viewing in radiology: seeing more, missing less? Advances in Health Sciences Education. 2016;21(1):189–205. 10.1007/s10459-015-9624-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28. Nodine CF, Kundel HL, Lauver SC, Toto LC. Nature of expertise in searching mammograms for breast masses. Academic radiology. 1996;3(12):1000–1006. 10.1016/S1076-6332(96)80032-8 [DOI] [PubMed] [Google Scholar]
  • 29.Castner N, Kasneci E, Kübler T, Scheiter K, Richter J, Eder T, et al. Scanpath comparison in medical image reading skills of dental students: distinguishing stages of expertise development. In: Proceedings of the 2018 ACM Symposium on Eye Tracking Research & Applications. ACM; 2018. p. 39.
  • 30. Veltman J, Gaillard A. Physiological workload reactions to increasing levels of task difficulty. Ergonomics. 1998;41(5):656–669. 10.1080/001401398186829 [DOI] [PubMed] [Google Scholar]
  • 31. Sweller J. Cognitive load during problem solving: Effects on learning. Cognitive science. 1988;12(2):257–285. 10.1207/s15516709cog1202_4 [DOI] [Google Scholar]
  • 32. Lance CE, Hedge JW, Alley WE. Joint relationships of task proficiency with aptitude, experience, and task difficulty: A cross-level, interactional study. Human Performance. 1989;2(4):249–272. 10.1207/s15327043hup0204_2 [DOI] [Google Scholar]
  • 33. Patel VL, Groen GJ. Knowledge based solution strategies in medical reasoning. Cognitive science. 1986;10(1):91–116. 10.1207/s15516709cog1001_4 [DOI] [Google Scholar]
  • 34. Patel VL, Groen GJ, Arocha JF. Medical expertise asa function of task difficulty. Memory & cognition. 1990;18(4):394–406. 10.3758/BF03197128 [DOI] [PubMed] [Google Scholar]
  • 35. Young JQ, Van Merrienboer J, Durning S, Ten Cate O. Cognitive load theory: implications for medical education: AMEE Guide No. 86. Medical teacher. 2014;36(5):371–384. 10.3109/0142159X.2014.889290 [DOI] [PubMed] [Google Scholar]
  • 36. Paas F, Ayres P. Cognitive load theory: A broader view on the role of memory in learning and education. Educational Psychology Review. 2014;26(2):191–195. 10.1007/s10648-014-9263-5 [DOI] [Google Scholar]
  • 37.Sibley C, Coyne J, Baldwin C. Pupil dilation as an index of learning. In: Proceedings of the Human Factors and Ergonomics Society Annual Meeting. vol. 55. SAGE Publications Sage CA: Los Angeles, CA; 2011. p. 237–241.
  • 38. Paas F, Tuovinen JE, Tabbers H, Van Gerven PW. Cognitive load measurement as a means to advance cognitive load theory. Educational psychologist. 2003;38(1):63–71. 10.1207/S15326985EP3801_8 [DOI] [Google Scholar]
  • 39.Appel T, Scharinger C, Gerjets P, Kasneci E. Cross-subject workload classification using pupil-related measures. In: Proceedings of the 2018 ACM Symposium on Eye Tracking Research & Applications. ACM; 2018. p. 4.
  • 40.Appel T, Sevcenko N, Wortha F, Tsarava K, Moeller K, Ninaus M, et al. Predicting Cognitive Load in an Emergency Simulation Based on Behavioral and Physiological Measures. In: 2019 International Conference on Multimodal Interaction; 2019. p. 154–163.
  • 41. Granholm E, Asarnow RF, Sarkin AJ, Dykes KL. Pupillary responses index cognitive resource limitations. Psychophysiology. 1996;33(4):457–461. 10.1111/j.1469-8986.1996.tb01071.x [DOI] [PubMed] [Google Scholar]
  • 42. Hyönä J, Tommola J, Alaja AM. Pupil dilation as a measure of processing load in simultaneous interpretation and other language tasks. The Quarterly Journal of Experimental Psychology. 1995;48(3):598–612. 10.1080/14640749508401407 [DOI] [PubMed] [Google Scholar]
  • 43. Beatty J. Task-evoked pupillary responses, processing load, and the structure of processing resources. Psychological bulletin. 1982;91(2):276 10.1037/0033-2909.91.2.276 [DOI] [PubMed] [Google Scholar]
  • 44. Unema PJA, Pannasch S, Joos M, Velichkovsky BM. Time course of information processing during scene perception: The relationship between saccade amplitude and fixation duration. Visual Cognition. 2005;12(3):473–494. 10.1080/13506280444000409 [DOI] [Google Scholar]
  • 45.Velichkovsky BM, Joos M, Helmert JR, Pannasch S. Two Visual Systems and their Eye Movements: Evidence from Static and Dynamic Scene Perception. In: CogSci 2005: Proceedings of the XXVII Conference of the Cognitive Science Society. Stresa, Italy; 2005. p. 2283–2288.
  • 46. Kok EM, De Bruin AB, Robben SG, Van Merrienboer JJ. Looking in the same manner but seeing it differently: Bottom-up and expertise effects in radiology. Applied Cognitive Psychology. 2012;26(6):854–862. 10.1002/acp.2886 [DOI] [Google Scholar]
  • 47. Kahneman D, Beatty J. Pupil diameter and load on memory. Science. 1966;154(3756):1583–1585. 10.1126/science.154.3756.1583 [DOI] [PubMed] [Google Scholar]
  • 48. Birren JE, Casperson RC, Botwinick J. Age changes in pupil size. Journal of Gerontology. 1950;5(3):216–221. 10.1093/geronj/5.3.216 [DOI] [PubMed] [Google Scholar]
  • 49. Brunyé TT, Eddy MD, Mercan E, Allison KH, Weaver DL, Elmore JG. Pupil diameter changes reflect difficulty and diagnostic accuracy during medical image interpretation. BMC medical informatics and decision making. 2016;16(1):77 10.1186/s12911-016-0322-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50. Szulewski A, Roth N, Howes D. The use of task-evoked pupillary response as an objective measure of cognitive load in novices and trained physicians: a new tool for the assessment of expertise. Academic Medicine. 2015;90(7):981–987. 10.1097/ACM.0000000000000677 [DOI] [PubMed] [Google Scholar]
  • 51. Van Gerven PW, Paas F, Van Merriënboer JJ, Schmidt HG. Memory load and the cognitive pupillary response in aging. Psychophysiology. 2004;41(2):167–174. 10.1111/j.1469-8986.2003.00148.x [DOI] [PubMed] [Google Scholar]
  • 52. Winn B, Whitaker D, Elliott DB, Phillips NJ. Factors affecting light-adapted pupil size in normal human subjects. Investigative ophthalmology & visual science. 1994;35(3):1132–1137. [PubMed] [Google Scholar]
  • 53. van der Wel P, van Steenbergen H. Pupil dilation as an index of effort in cognitive control tasks: A review. Psychonomic bulletin & review. 2018;25(6):2005–2015. 10.3758/s13423-018-1432-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54. Porter G, Troscianko T, Gilchrist ID. Effort during visual search and counting: Insights from pupillometry. The Quarterly Journal of Experimental Psychology. 2007;60(2):211–229. 10.1080/17470210600673818 [DOI] [PubMed] [Google Scholar]
  • 55. Backs RW, Walrath LC. Eye movement and pupillary response indices of mental workload during visual search of symbolic displays. Applied ergonomics. 1992;23(4):243–254. 10.1016/0003-6870(92)90152-L [DOI] [PubMed] [Google Scholar]
  • 56. Geng JJ, Blumenfeld Z, Tyson TL, Minzenberg MJ. Pupil diameter reflects uncertainty in attentional selection during visual search. Frontiers in human neuroscience. 2015;9:435 10.3389/fnhum.2015.00435 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 57. Takeuchi T, Puntous T, Tuladhar A, Yoshimoto S, Shirama A. Estimation of mental effort in learning visual search by measuring pupil response. PloS one. 2011;6(7):e21973 10.1371/journal.pone.0021973 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58. Lowenstein O, Feinberg R, Loewenfeld IE. Pupillary movements during acute and chronic fatigue: A new test for the objective evaluation of tiredness. Investigative Ophthalmology & Visual Science. 1963;2(2):138–157. [Google Scholar]
  • 59. Murata A. Assessment of fatigue by pupillary response. IEICE TRANSACTIONS on Fundamentals of Electronics, Communications and Computer Sciences. 1997;80(7):1318–1323. [Google Scholar]
  • 60. Abokyi S, Owusu-Mensah J, Osei K. Caffeine intake is associated with pupil dilation and enhanced accommodation. Eye. 2017;31(4):615 10.1038/eye.2016.288 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 61. Bradley MM, Miccoli L, Escrig MA, Lang PJ. The pupil as a measure of emotional arousal and autonomic activation. Psychophysiology. 2008;45(4):602–607. 10.1111/j.1469-8986.2008.00654.x [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 62. Gerjets P, Scheiter K, Catrambone R. Designing instructional examples to reduce intrinsic cognitive load: Molar versus modular presentation of solution procedures. Instructional Science. 2004;32(1-2):33–58. 10.1023/B:TRUC.0000021809.10236.71 [DOI] [Google Scholar]
  • 63. Mathôt S, Fabius J, Van Heusden E, Van der Stigchel S. Safe and sensible preprocessing and baseline correction of pupil-size data. Behavior research methods. 2018;50(1):94–106. 10.3758/s13428-017-1007-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 64. Zheng B, Jiang X, Atkins MS. Detection of changes in surgical difficulty: evidence from pupil responses. Surgical innovation. 2015;22(6):629–635. 10.1177/1553350615573582 [DOI] [PubMed] [Google Scholar]
  • 65. Tien T, Pucher PH, Sodergren MH, Sriskandarajah K, Yang GZ, Darzi A. Differences in gaze behaviour of expert and junior surgeons performing open inguinal hernia repair. Surgical endoscopy. 2015;29(2):405–413. 10.1007/s00464-014-3683-7 [DOI] [PubMed] [Google Scholar]
  • 66. Szulewski A, Kelton D, Howes D. Pupillometry as a Tool to Study Expertise in Medicine. Frontline Learning Research. 2017;5(3):53–63. 10.14786/flr.v5i3.256 [DOI] [Google Scholar]
  • 67. Brunyé TT, Drew T, Weaver DL, Elmore JG. A review of eye tracking for understanding and improving diagnostic interpretation. Cognitive research: principles and implications. 2019;4(1):7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 68. Turgeon DP, Lam EW. Influence of experience and training on dental students’ examination performance regarding panoramic images. Journal of dental education. 2016;80(2):156–164. [PubMed] [Google Scholar]
  • 69. Castner N, Kübler TC, Richter J, Eder T, Huettig F, Keutel C, et al. Deep semantic gaze embedding and scanpath comparison for expertise classification during OPT viewing In: Eye Tracking Research and Applications. ACM; 2020. [Google Scholar]
  • 70. Nodine CF, Kundel HL, Mello-Thoms C, Weinstein SP, Orel SG, Sullivan DC, et al. How experience and training influence mammography expertise. Academic radiology. 1999;6(10):575–585. 10.1016/S1076-6332(99)80252-9 [DOI] [PubMed] [Google Scholar]
  • 71. Manning D, Ethell S, Donovan T, Crawford T. How do radiologists do it? The influence of experience and training on searching for chest nodules. Radiography. 2006;12(2):134–142. 10.1016/j.radi.2005.02.003 [DOI] [Google Scholar]
  • 72. Krupinski EA, Tillack AA, Richter L, Henderson JT, Bhattacharyya AK, Scott KM, et al. Eye-movement study and human performance using telepathology virtual slides. Implications for medical education and differences with experience. Human pathology. 2006;37(12):1543–1556. 10.1016/j.humpath.2006.08.024 [DOI] [PubMed] [Google Scholar]
  • 73. American College of Radiology. ACR technical standard for digital image data management. Reston (VA): American College of Radiology; 2002; p. 811–9. [Google Scholar]
  • 74. Brennan PC, McEntee M, Evanoff M, Phillips P, O’Connor WT, Manning DJ. Ambient lighting: effect of illumination on soft-copy viewing of radiographs of the wrist. American Journal of Roentgenology. 2007;188(2):W177–W180. 10.2214/AJR.05.2048 [DOI] [PubMed] [Google Scholar]
  • 75. Goo JM, Choi JY, Im JG, Lee HJ, Chung MJ, Han D, et al. Effect of monitor luminance and ambient light on observer performance in soft-copy reading of digital chest radiographs. Radiology. 2004;232(3):762–766. 10.1148/radiol.2323030628 [DOI] [PubMed] [Google Scholar]
  • 76. Pollard BJ, Samei E, Chawla AS, Beam C, Heyneman LE, Koweek LMH, et al. The effects of ambient lighting in chest radiology reading rooms. Journal of digital imaging. 2012;25(4):520–526. 10.1007/s10278-012-9459-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 77. Kagadis GC, Walz-Flannigan A, Krupinski EA, Nagy PG, Katsanos K, Diamantopoulos A, et al. Medical imaging displays and their use in image interpretation. Radiographics. 2013;33(1):275–290. 10.1148/rg.331125096 [DOI] [PubMed] [Google Scholar]
  • 78.Winkler T. NotebookCheck Review Dell Precision M4800 Notebook; 2013. Available from: https://www.notebookcheck.net/Review-Dell-Precision-M4800-Notebook.104416.0.html.
  • 79.Ngo A. NotebookCheck Review HP ZBook 15 Workstation; 2014. Available from: https://www.notebookcheck.net/Review-HP-ZBook-15-Workstation.108229.0.html.
  • 80.SensoMotoric Instruments. BeGaze Manual; 2017.
  • 81. Hoeks B, Levelt WJ. Pupillary dilation as a measure of attention: A quantitative system analysis. Behavior Research Methods, Instruments, & Computers. 1993;25(1):16–26. 10.3758/BF03204445 [DOI] [Google Scholar]
  • 82.Klingner J, Kumar R, Hanrahan P. Measuring the task-evoked pupillary response with a remote eye tracker. In: Proceedings of the 2008 symposium on Eye tracking research & applications. ACM; 2008. p. 69–72.
  • 83.Kiefer P, Giannopoulos I, Duchowski A, Raubal M. Measuring cognitive load for map tasks through pupil diameter. In: The Annual International Conference on Geographic Information Science. Springer; 2016. p. 323–337.
  • 84. Brunyé TT, Gardony AL. Eye tracking measures of uncertainty during perceptual decision making. International Journal of Psychophysiology. 2017;120:60–68. 10.1016/j.ijpsycho.2017.07.008 [DOI] [PubMed] [Google Scholar]
  • 85. Ahern S, Beatty J. Pupillary responses during information processing vary with Scholastic Aptitude Test scores. Science. 1979;205(4412):1289–1292. 10.1126/science.472746 [DOI] [PubMed] [Google Scholar]
  • 86. Verney SP, Granholm E, Marshall SP. Pupillary responses on the visual backward masking task reflect general cognitive ability. International Journal of Psychophysiology. 2004;52(1):23–36. 10.1016/j.ijpsycho.2003.12.003 [DOI] [PubMed] [Google Scholar]
  • 87. Chen S, Epps J. Using Task-Induced Pupil Diameter and Blink Rate to Infer Cognitive Load. Human–Computer Interaction. 2014;29(4):390–413. 10.1080/07370024.2014.892428 [DOI] [Google Scholar]
  • 88.Duchowski AT, Krejtz K, Krejtz I, Biele C, Niedzielska A, Kiefer P, et al. The Index of Pupillary Activity: Measuring Cognitive Load Vis-à-vis Task Difficulty with Pupil Oscillation. In: Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems. CHI’18. New York, NY: ACM; 2018. p. 282:1–282:13. Available from: http://doi.acm.org/10.1145/3173574.3173856.
  • 89. Duchowski A, Krejtz K, Żurawska J, House D. Using Microsaccades to Estimate Task Difficulty During Visual Search of Layered Surfaces. IEEE Transactions on Visualization and Computer Graphics. 2019;213. [DOI] [PubMed] [Google Scholar]
  • 90. Kundel HL, Nodine CF, Carmody D. Visual scanning, pattern recognition and decision-making in pulmonary nodule detection. Investigative radiology. 1978;13(3):175–181. 10.1097/00004424-197805000-00001 [DOI] [PubMed] [Google Scholar]
  • 91. Kundel HL, Nodine CF, Conant EF, Weinstein SP. Holistic component of image perception in mammogram interpretation: gaze-tracking study. Radiology. 2007;242(2):396–402. 10.1148/radiol.2422051997 [DOI] [PubMed] [Google Scholar]
  • 92. Kundel HL, Nodine CF, Krupinski EA, Mello-Thoms C. Using gaze-tracking data and mixture distribution analysis to support a holistic model for the detection of cancers on mammograms. Academic radiology. 2008;15(7):881–886. 10.1016/j.acra.2008.01.023 [DOI] [PubMed] [Google Scholar]
  • 93. Attar N, Schneps MH, Pomplun M. Working memory load predicts visual search efficiency: Evidence from a novel pupillary response paradigm. Memory & cognition. 2016;44(7):1038–1049. 10.3758/s13421-016-0617-8 [DOI] [PMC free article] [PubMed] [Google Scholar]

Decision Letter 0

Susana Martinez-Conde

23 Dec 2019

PONE-D-19-27421

Pupil diameter differentiates expertise in dental radiography visual search

PLOS ONE

Dear Ms. Castner,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please address the methodological concerns raised by both reviewers. Your resubmission should also improve the presentation of results and overall manuscript structure, as delineated by Reviewer #2. 

We would appreciate receiving your revised manuscript by Feb 04 2020 11:59PM. When you are ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter.

To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). This letter should be uploaded as separate file and labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. This file should be uploaded as separate file and labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. This file should be uploaded as separate file and labeled 'Manuscript'.

Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out.

We look forward to receiving your revised manuscript.

Kind regards,

Susana Martinez-Conde

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

http://www.journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and http://www.journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Your ethics statement must appear in the Methods section of your manuscript. If your ethics statement is written in any section besides the Methods, please move it to the Methods section and delete it from any other section. Please also ensure that your ethics statement is included in your manuscript, as the ethics section of your online submission will not be published alongside your manuscript.

3. We note that Figure 2 includes an image of participants in the study.

As per the PLOS ONE policy (http://journals.plos.org/plosone/s/submission-guidelines#loc-human-subjects-research) on papers that include identifying, or potentially identifying, information, the individual(s) or parent(s)/guardian(s) must be informed of the terms of the PLOS open-access (CC-BY) license and provide specific permission for publication of these details under the terms of this license. Please download the Consent Form for Publication in a PLOS Journal (http://journals.plos.org/plosone/s/file?id=8ce6/plos-consent-form-english.pdf). The signed consent form should not be submitted with the manuscript, but should be securely filed in the individual's case notes. Please amend the methods section and ethics statement of the manuscript to explicitly state that the patient/participant has provided consent for publication: “The individual in this manuscript has given written informed consent (as outlined in PLOS consent form) to publish these case details”.

If you are unable to obtain consent from the subject of the photograph, you will need to remove the figure and any other textual identifying information or case descriptions for this individual.

4. We note you have included a table to which you do not refer in the text of your manuscript. Please ensure that you refer to Table 2 in your text; if accepted, production will need this reference to link the reader to the Table.

5. In your Data Availability statement, you have not specified where the minimal data set underlying the results described in your manuscript can be found; the address provided (ftp://peg-public:peg-public@messor.informatik.uni-tuebingen.de/norac/VisualExpertiseRadiology.zip) is not found.

PLOS defines a study's minimal data set as the underlying data used to reach the conclusions drawn in the manuscript and any additional data required to replicate the reported study findings in their entirety. All PLOS journals require that the minimal data set be made fully available. For more information about our data policy, please see http://journals.plos.org/plosone/s/data-availability.

Upon re-submitting your revised manuscript, please upload your study’s minimal underlying data set as either Supporting Information files or to a stable, public repository and include the relevant URLs, DOIs, or accession numbers within your revised cover letter. For a list of acceptable repositories, please see http://journals.plos.org/plosone/s/data-availability#loc-recommended-repositories. Any potentially identifying patient information must be fully anonymized.

Important: If there are ethical or legal restrictions to sharing your data publicly, please explain these restrictions in detail. Please see our guidelines for more information on what we consider unacceptable restrictions to publicly sharing data: http://journals.plos.org/plosone/s/data-availability#loc-unacceptable-data-access-restrictions. Note that it is not acceptable for the authors to be the sole named individuals responsible for ensuring data access.

We will update your Data Availability statement to reflect the information you provide in your cover letter.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: No

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: This study investigated pupil-dilation markers of expertise in dental radiography visual search. This including the effect of inspecting anomalies of different difficulty. The authors found that during search for anomalies, experts had smaller pupil size change from baseline, indicating lower cognitive load as expected. In addition, the experts’, but not the students’ pupil dilation increased with anomaly discrimination difficulty, possibly because the students could not tell the difference.

As far as I can tell, these results have not been published elsewhere. However, I think that they do not crate any shift in thinking or make a big difference from previous studies (e.g. of radiology).

In general, the study conforms with appropriate standards (experiments, analyses, stats) and are described in great details (sometimes too long in my opinion), although some improvements could be made in data presentation (see below).

The conclusions are appropriate and supported by the data.

The research meets ethics standards, and all data are claimed to be available.

Regarding the presentation:

1. The writing is too detailed and sometimes unnecessary long in my view. One example that stands out is the “conclusions” section. This is a long text, which belongs (if needed) to the discussion, after proper shortening. It should be replaced by a short Conclusions par.

2. All group data are presented as pupil size change, except from figure 1 where the absolute pupil is presented, showing a large difference between the experts and the others, and generally smaller pupil with experience (sixth is larger than the later years). Why is this figure (1) only mentioned in Methods? Why not present the data as averages with error bars? It looks like some of the results do not conform to the expected order based on student experience (better show) – does a relative measure conform with the expected order? Also, I understand that the variability of the absolute pupil size is large and perhaps this is why no error bars are shown in fig 1, but worth seeing these effects in some way. Also, why not presenting the pupil size change in %?

3. Only group averages are shown (other than examples). It would be useful to show scatter plots of individual data, e.g pupil size change vs pupil size for the different groups.

4. Intro, 1st line: Mental imagery is a commonly performed task – is this the right phrase?

Reviewer #2: I think this is an interesting study, worthy of publication. The manuscript, however, needs to be reorganized because it lacks structure and consistency. Moreover, is not in line with the Plos One guidelines. I would be happy to discuss smaller issues – in a next round of reviewing – once the structural issues of the manuscript (see below) are fixed.

The introduction: according to PlosOne guidelines, should include a brief review of the key literature. It, however, counts more than five pages spread over four different sections (Introduction, Background Characterizing Expertise, Eye Movement Behavior Reflective of Cognitive Processes, and Related Work). I suggest limiting the introduction to 2-3 pages, and focusing on review that is most relevant for this study. Parts of the “Related Work” section could be recycled for the discussion section.

The method section: Overall, the method section could be written more concise and better organized. Now it is chaotic and lacks structure, consistency and order. I suggest organizing the method sections into the following subsections:

1. Participants: which should be limited to explaining how many students / experts participated in the study and how they were recruited and/or selected (referring to Table 1). Note that summer and winter semesters are uncommon in the US, as well as to express their academic career stage in terms of semesters. So maybe you could phrase it in a way that is understandable for both European and American readers. I think the explanation of figure 1, and the figure itself should be part of the result section.

2. Experimental paradigm: the information currently provided in the section “data acquisition” and shown in Figure 3.

3. Stimuli: describe the OPTs, anomalies, provide example of the anomalies, and explain how the anomaly ground truths were generated; and that maps are generated based upon the ground truths [Current figure 6 and 7]

4. Data acquisition: mention here the specs of the eye tracker and laptop, and briefly describe “the environment”. There is a danger that readers will perceive the “simultaneous” data collecting (homogeneous circumstances) as a strategy to reduce noise (which would of have been true if the data from both students and experts were collected at the same time, in the same room, and under the same circumstances). Since the data collection, however, happened simultaneously for each group separately, small differences, not controlled for, might be introduced between the two groups. The benefits (in this study) of simultaneous data collection should thus be downplayed.

5. Data Preprocessing and data analysis: describe here, concise, the criteria for selecting raw data and excluding data (blinks); how the pupil diameter was measured, etc…, and how gaze combined with the map of the ground truths yields gaze hits, etc. Minor: you mention that 199 datasets were excluded, but it will be more informative if you also report how many sets were collected in the first place.

Additional concerns, related to the method section

1. Because the exploration phase was shorter (45 vs 90 seconds) for experts than students, and the session counted fewer OPTs (15 vs 20) for experts than students, the overall duration of a session was significantly longer for students than for experts. There is thus a concern that fatigue (exhibited by the students) might of have influenced the results. Whether fatigue plays a role, or not, can easily be addressed: instead of looking for differences between students and experts (Figures 8 and 9), the same analysis can be performed to check for differences between behavior before (first set of 10 OPTs) and after (second set of 10 OPTs) the break. I encourage the authors to perform this extra analysis, and to report the outcome.

2. (Minor) Neither in the body text nor in the accompanying figure it is explained what the difference is between the read and the green dots (Figure 4a). Is it blinks for respectively left and right eye? Is so, please mention.

3. Students evaluated 20 OPTs. However, according the manuscript, only maps were created for 15 OPTs. Please clarify.

Results: I suggest starting the result section by explaining the current figure 3. It nicely shows that pupil diameter decreases with expertise; therefore, it can be considered a result. Moreover, it allows a nice and smooth transition to the results shown in figures 8 and 9.

Discussion: There is quiet some redundancy is this section. I suggest to (1) write it more concise, (2) transfer information from “Conclusion” to “Discussion” (see my next comment); and (3) optionally, use some of the information – now in the section “Related Work” – to compare your current results with those of previous studies.

Conclusion: There is indeed a concern that the study was not controlled for age. That discussion, however, does not belong in the section “Conclusion” and should be moved to the section “Discussion”. This also applies to the discussion of technical problems associated with the eye tracker data collection. Scholars who do not have the time to read the whole article will likely only read the abstract and conclusion. So I suggest to state clearly, and in sufficient detail, the main findings here.

Figures: I do not think that there is a need for nine figures, I suggest to reorganize them according to topic. Here is a suggestion for figures:

- One figure for acquisition:

o New Figure 1, panel A: the current Figure 3

o New Figure 1, panel B: the current Figure 2

- One figure for signal processing

o New Figure 2, panels A and B: the current Figure 4

o New Figure 2, panel C: the current Figure 5

- One figure for ground truths and maps

o New Figure 3: the current Figure 6, but replace panel D (not informative) with Figure 7

- One figure for results

o New Figure 4, panel A: current Figure 1

o New Figure 4, panel B: current Figure 8

o New Figure 4, panel C: current Figure 9

Other concerns: No reference is made, in the text, to Table 2. Please check.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: Yes: Nicolas Brunet

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files to be viewed.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2020 May 29;15(5):e0223941. doi: 10.1371/journal.pone.0223941.r002

Author response to Decision Letter 0


2 Feb 2020

Dear Dr. Susana Martinez-Conde,

We would like to thank you and the reviewers for the constructive feedback. We appreciate that we are given the chance to revise and resubmit another version of the manuscript based on the reviewers’ concerns. In the following, we address each of the comments and report on how the manuscript has been changed accordingly.

Editor Summary:

*Please address the methodological concerns raised by both reviewers, the presentation of results, and the overall structure regarding the manuscript.

We have made major changes to our manuscript regarding these issues as well as conformed it better to the PLOS ONE style requirements. Please find more specific comments regarding these issues as responses to the respective reviewer remarks. We hope these changes have been done in a satisfactory manner.

*Your ethics statement must appear in the Methods section of your manuscript.

The appropriate change to the methods has been added to lines 185-195.

*PLOS-ONE policy regarding consent from the participants pictured in Figure 2.

As we only required verbal consent from the students in the image, we have removed figure 2. The only face visible from this angle, is that from the instructor. However, if it is determined that verbal consent and no shown participant faces is still appropriate, we can add Figure 2 back into the manuscript.

*We note that Table 2 is not referred to in the text of your manuscript.

We have now included a reference in our results to the Table in the supporting information. See line 482.

*The address provided that makes the data publicly available is not found.

The data can be found now at this new link:

ftp://peg-public:peg-public@messor.informatik.uni-tuebingen.de/peg-public/norac/VisualExpertiseRadiology.zip

Reviewer #1:

We thank you for your contributions towards the review and hope our changes are suitable address your comments (*).

*As far as I can tell, these results have not been published elsewhere. However, I think they do not create any shift in thinking or make a big difference from previous studies (e.g. Radiology).

Previous research, we feel, has only scratched the surface of the cognitive processes during visual inspection of radiological images and the dichotomy between experts and novices. For instance, many of the papers cited in the related work measure cognitive load indicators during the whole visual inspection or the whole task. Accordingly, they have one data point for one participant that is supposed to represent their whole decision-making process during that task. However, one value is not enough to understand this process for such complex stimuli such as radiographs. For this reason, our work goes one step further by looking at cognitive processes on the sub-region level during visual inspection and gathers multiple measures for a participant that are more informative of the regions he or she is inspecting.

We understand that in our manuscript, we may have not stressed the deeper level of investigation enough. Therefore, the appropriate changes were made to the introduction (lines 41-45,129-130,148-150) and strengthen general discussion and conclusion.

*The writing is too detailed and sometimes unnecessary long in my view. One example that stands out is the “conclusions” section. This is a long text, which belongs (if needed) to the discussion, after proper shortening. It should be replaced by a short Conclusions paragraph.

We agree that we got carried away with the details sometimes leading to unnecessarily long parts. Specifically, we have now shortened the conclusion substantially (2 short paragraphs starting at line 495) by incorporating portions into the discussion where it was deemed more appropriate. Regarding general lengthiness, we condensed the introduction and restricted the methods and results as per comments from reviewer #2 and general concerns from the editor.

*All group data are presented as pupil size change, except from figure 1 where the absolute pupil is presented, showing a large difference between the experts and the others, and generally smaller pupil with experience (sixth is larger than the later years). Why is this figure (1) only mentioned in Methods? Why not present the data as averages with error bars? It looks like some of the results do not conform to the expected order based on student experience (better show) – does a relative measure conform with the expected order? Also, I understand that the variability of the absolute pupil size is large and perhaps this is why no error bars are shown in fig 1, but worth seeing these effects in some way.

We initially chose to show figure 1 in the participants part of the methods to support why we chose only one group of students for comparison to experts. Reviewer #2 also pointed out problems with figure 1, and you can see our comments. However, we thank you for pointing out that this graph does not fully fit with the rest of the work.

Therefore, we have removed figure 1 to avoid further confusion.

*Also, why not presenting the pupil size change in %?

Upon feedback from authors AD and TA as well as further research, it was determined that absolute pupil size change was the more optimal measurement since it is more sensitive, though robust to spurious effects.

In [Mahot et al, 2018], they argue that subtractive has higher statistical power than divisive (pupil size/baseline).

[Mathôt, S., Fabius, J., Van Heusden, E., & Van der Stigchel, S. (2018). Safe and sensible preprocessing and baseline correction of pupil-size data. Behavior research methods, 50(1), 94–106. doi:10.3758/s13428-017-1007-2]

We justify our choice for absolute change from baseline (subtractive) metric in lines 347-349 and added the above source.

*Only group averages are shown (other than examples). It would be useful to show scatter plots of individual data, e.g pupil size change vs pupil size for the different groups.

For the sake of brevity and readability, we chose to show group differences via the averages. As we have over 500 data sets, we felt scatter plots would be harder for the readers to interpret. The data is however publicly available for further visualization purposes.

*Intro, 1st line: Mental imagery is a commonly performed task – is this the right phrase?

We rewrote the sentence (line 2) and changed mental imagery to visual inspection to be more understandable.

Reviewer #2:

We thank you for the positive comments regarding the research and are thankful that you spent the time preparing such concrete suggestions (*).

*The introduction should include a brief review of the key literature. I suggest limiting the introduction to 2-3 pages, and focusing on review that is most relevant for this study. Parts of the “Related Work” section could be recycled for the discussion section.

We have limited the introduction section to 3 pages. It has been reorganized as follows:

- Introduction

o Background characterizing expertise

-- Expertise and memory

-- Skill acquisition and cognitive load

o Pupil Diameter as a measure for cognitive load

-- In Visual search

-- Factors affecting pupillary response

Multiple sentences, paragraphs, and a section have been removed after determining that they did not significantly contribute to the overall narrative. Related work as its own section is on the fourth page and has been highly summarized (2 paragraphs now), and references are further detailed in the discussion section.

*Overall, the method section could be written more concise and better organized. Now it is chaotic and lacks structure, consistency and order. I suggest organizing the method sections into the following subsections:

*1. Participants: which should be limited to explaining how many students / experts participated in the study and how they were recruited and/or selected (referring to Table 1). *Note that summer and winter semesters are uncommon in the US, as well as to express their academic career stage in terms of semesters. So maybe you could phrase it in a way that is understandable for both European and American readers. I think the explanation of figure 1, and the figure itself should be part of the result section.

Since our analysis focuses on a subgroup of students collected as part of a much larger study, we felt it was necessary to support our subgroup choice. We had placed this information as well as figure 1 in the participants section of the methods to be more transparent of who we chose to analyze. We did not run each group against experts, rather, we assessed the subgroup of the sixth semester students to be the most representative of high cognitive load. We have removed figure 1 to avoid any further confusion. We hope we have clarified this in lines 165, 172-175.

We do agree with you that the semester distinction is not common other regions. Therefore, we made changes to lines 167-169 where we specify which year the student is in more understandable. We also changed line 165 to say data collected over multiple semesters, to avoid the summer winter confusion, since the summer semester would be equivalent to the spring semester and the winter semester would be equivalent to the fall semester, though this is not pertinent to the overall methods.

*Experimental paradigm: the information currently provided in the section “data acquisition” and shown in Figure 3.

We renamed subsection 2 to Experimental paradigm and this section has all the information that was previously found in the data collection section and includes a Figure 2 (previously figure 3), which is now the outline of the experimental section.

*Stimuli: describe the OPTs, anomalies, provide example of the anomalies, and explain how the anomaly ground truths were generated; and that maps are generated based upon the ground truths [Current figure 6 and 7]

We renamed section 3 to Stimuli and we further break this section down into 2 parts: 1) OPT Content (e.g. examples, types of anomalies, ground truth generation) and reference figure 3 (previously figure 6) and 2) Ground Truth Maps and reference figure 3.D (previously figure 7) .

*Data acquisition: mention here the specs of the eye tracker and laptop, and briefly describe “the environment”. There is a danger that readers will perceive the “simultaneous” data collecting (homogeneous circumstances) as a strategy to reduce noise (which would of have been true if the data from both students and experts were collected at the same time, in the same room, and under the same circumstances). Since the data collection, however, happened simultaneously for each group separately, small differences, not controlled for, might be introduced between the two groups. The benefits (in this study) of simultaneous data collection should thus be downplayed.

We created a section called Data acquisition and broke it further down into subsections detailing 1) the environment, 2) laptop, and 3) eye tracker.

We agree that the simultaneous data collection should be downplayed to reduce confusion from the reader. Therefore, we removed sentences in the environment part of the data acquisition subsection highlighting the benefits. Additionally, as per issues with consent, we have also removed Figure 2. The sentences related to figure 2 and the classroom have been shortened or removed.

*Data Preprocessing and data analysis: describe here, concise, the criteria for selecting raw data and excluding data (blinks); how the pupil diameter was measured, etc…, and how gaze combined with the map of the ground truths yields gaze hits, etc. Minor: you mention that 199 datasets were excluded, but it will be more informative if you also report how many sets were collected in the first place.

We have created a subsection called Data preprocessing. Here we subsection it to 1) raw data selection (e.g. only high tracking ratio/enough samples, baseline, trimming beginning and end of stim presentation) 2) blink and saccade removal 3) how pupil diameter measured 4) mapping to get gaze hits.

We have also added the total number of data set collected in the first place (see lines 311-312 and table 1).

*Because the exploration phase was shorter (45 vs 90 seconds) for experts than students, and the session counted fewer OPTs (15 vs 20) for experts than students, the overall duration of a session was significantly longer for students than for experts. There is thus a concern that fatigue (exhibited by the students) might of have influenced the results. Whether fatigue plays a role, or not, can easily be addressed: instead of looking for differences between students and experts (Figures 8 and 9), the same analysis can be performed to check for differences between behavior before (first set of 10 OPTs) and after (second set of 10 OPTs) the break. I encourage the authors to perform this extra analysis, and to report the outcome.

We have performed this additional analysis and found no effect attributable to fatigue between first set and second set of images for both students and experts. We have accordingly added the graph as Fig 1. in the Supporting Information and mention it in the results at lines 362-364. We can also add it directly to the results if you would prefer.

*Neither in the body text nor in the accompanying figure it is explained what the difference is between the read and the green dots (Figure 4a). Is it blinks for respectively left and right eye? Is so, please mention.

Indeed, the red and green dots are the time points when blinks are detected for the left and right eye. We have updated the caption in figure 4a to clarify this and have added a legend in both figure 4a and 4b.

*Students evaluated 20 OPTs. However, according the manuscript, only maps were created for 15 OPTs. Please clarify.

Due to time constraints, experts only viewed 15 of those 20 OPTs. Therefore, we only analyzed students’ data for those same 15 OPTs that the experts viewed. We wanted to control for any possible effects on the student data for images unseen by the experts. We state this more appropriately in lines 219-220.

*Results: I suggest starting the result section by explaining the current figure 3. It nicely shows that pupil diameter decreases with expertise; therefore, it can be considered a result. Moreover, it allows a nice and smooth transition to the results shown in figures 8 and 9.

Referring to your concern on the participants, we feel that the original figure 1 is not really acceptable for a result, rather as a choice justification. Reviewer 1 pointed out how this graph does not actually fit with the rest of the paper, and we do agree and have changed it accordingly. We also feel we cannot state there is a decrease with knowledge gain in students as we have not performed this analysis, due to the amount and size of the raw data files.

In general, we rewrote to results section as per feedback from one of our authors.

*Discussion: There is quiet some redundancy is this section. I suggest to (1) write it more concise, (2) transfer information from “Conclusion” to “Discussion” (see my next comment); and (3) optionally, use some of the information – now in the section “Related Work” – to compare your current results with those of previous studies.

We have done some substantial revisions to the discussion to remove redundancy, incorporate most of the conclusion, address the age control, limitations with the eye tracker, and we elaborate on the sources from the related work and compare our results to them.

*Conclusion: There is indeed a concern that the study was not controlled for age. That discussion, however, does not belong in the section “Conclusion” and should be moved to the section “Discussion”. This also applies to the discussion of technical problems associated with the eye tracker data collection. Scholars who do not have the time to read the whole article will likely only read the abstract and conclusion. So I suggest to state clearly, and in sufficient detail, the main findings here.

We substantially rewrote the conclusion to be shorter and detail only the main findings.

*Figures: I do not think that there is a need for nine figures, I suggest to reorganize them according to topic.

We reorganized the figures as follows:

- New Figure 1 (original removed), Old Figure 3

- New Figure 2, Old Figures 6 & 7

- Figure 3 and 4, Originally 4 and 5

- New Figure 5, Old Figures 8 & 9

Attachment

Submitted filename: Response to Reviewers.docx

Decision Letter 1

Susana Martinez-Conde

26 Mar 2020

PONE-D-19-27421R1

Pupil diameter differentiates expertise in dental radiography visual search

PLOS ONE

Dear Ms. Castner,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

==============================

Reviewer 2 has suggeted a few minor revisions and edits that should be considered and addressed. 

==============================

We would appreciate receiving your revised manuscript by May 10 2020 11:59PM. When you are ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter.

To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). This letter should be uploaded as separate file and labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. This file should be uploaded as separate file and labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. This file should be uploaded as separate file and labeled 'Manuscript'.

Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out.

We look forward to receiving your revised manuscript.

Kind regards,

Susana Martinez-Conde

Academic Editor

PLOS ONE

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

Reviewer #2: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: (No Response)

Reviewer #2: The authors have addressed my concerns; the manuscript is now much more structured and hence a lot easier to read and understand. Here follows a list with minor comments.

LINES 21/22: Since the abbreviation “OPT” is not intuitive, “…inspectors of dental panoramic radiographs (OPT)” might suggest that OPT stands for specialists examining radiographs (rather than for the radiographs themselves).

Traditionally the “Introduction” section comes after “Abstract” and before “Material and Methods”. The manuscript, however, contains additional (non-standard) sections (“Background” and “Related work”), which might violate the submission guidelines. I suggest using subheadings, so that all information provided between lines 2 and 162 falls under “Introduction”.

There is still significant overlap/redundancy between paragraph 35-45 and paragraph 152-162.

LINE 71: “… and devote shorter fixation time to an anomaly”. Is this correct? From the provided references, I understand that the experts show a fast initial fixation on the abnormality; do they also devote shorter fixation time to an anomaly than novices?

LINE 205-206: Please revise the following sentence: “There was unlimited time for the marking phase, and continued with a button click.”

LINE 227: “shows four OPT images viewed in the experiment”; should be “shows three OPT….”

LINE 246-247: Please revise the following sentence: “Certain anomalies inside another and that were highly similar in nature,…”

LINE 252: “… an spatial accuracy errors in the gaze.” Should be “… the spatial accuracy errors in the gaze.”

Caption figure 3: “… could incorrectly labeled as blinks. “. Something about the construction of this sentence is not right.

LINES 329-330: “Since baseline data was the two second fixation cross directly before each stimulus…”. Do you mean: “Since baseline data was sampled during the two seconds the fixation cross was displayed….”?

LINES 445-446: Please revise the following sentence: “Whereas bottom-up strategies that student generally employ focus on salient, noticeable images features regardless of relevancy and is less efficient.”

LINES 499-501: Please revise the following sentence: “Experts, renowned for their streamlined processing abilities, are able to selectively allocate their attention to relevant information and is evident in the pupillary response.”

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: Yes: Nicolas Brunet

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files to be viewed.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2020 May 29;15(5):e0223941. doi: 10.1371/journal.pone.0223941.r004

Author response to Decision Letter 1


7 May 2020

Dear Dr. Susana Martinez-Conde,

We would like to thank the reviewers for their constructive feedback and suggestions to further improve the manuscript. We are pleased that reviewer #1 feels that all comments have been addressed. In the following, please find reviewer #2’s suggestions (*) and the changes made to cover these.

Response to Reviewer #2 Comments:

Thanks for raising these points, we hope we have addressed them appropriately.

*LINES 21/22: Since the abbreviation “OPT” is not intuitive, “…inspectors of dental panoramic radiographs (OPT)” might suggest that OPT stands for specialists examining radiographs (rather than for the radiographs themselves).

We thank you for pointing out the clarity issue from the grammar. We have since split the sentence into two to avoid the sentence subject confusion. We have also added the technical term for these dental radiographs, which is where the abbreviation comes from.

*Traditionally the “Introduction” section comes after “Abstract” and before “Material and Methods”. The manuscript, however, contains additional (non-standard) sections (“Background” and “Related work”), which might violate the submission guidelines. I suggest using subheadings, so that all information provided between lines 2 and 162 falls under “Introduction”.

We would not like to violate the submission guidelines and we have made the background and related work subheadings to the introduction section. Now the introduction consists of three subsections: Background, cognitive load, related work. Paragraph headings in the introduction have also been removed, to avoid confusion as well.

Additionally, we realized that Plos One does not permit footnotes, thus we have removed all footnotes, only incorporating necessary ones into the text.

*There is still significant overlap/redundancy between paragraph 35-45 and paragraph 152-162.

We have removed the paragraph originally at lines 35-45 and combined a few sentences into the paragraph originally at lines 152-162.

*LINE 71: “… and devote shorter fixation time to an anomaly”. Is this correct? From the provided references, I understand that the experts show a fast initial fixation on the abnormality; do they also devote shorter fixation time to an anomaly than novices?

In the reference by van der Gijp and colleagues (2017), they found that for detection tasks, experts spent less time fixating on lesions compared to novices for detections tasks. However, they also point out that when the task involves diagnostic reasoning, then the fixation duration is higher than novices.

We have now specified in line 71 (now line 62) that shorter fixation durations for detection tasks.

*LINE 205-206: Please revise the following sentence: “There was unlimited time for the marking phase, and continued with a button click.”

We have changed the sentence to “There was unlimited time for the marking phase and participants could continue with a button click.”

*LINE 227: “shows four OPT images viewed in the experiment”; should be “shows three OPT….”

We have change “four” to “three”. Thank you for pointing out this inconsistency from the original version.

*LINE 246-247: Please revise the following sentence: “Certain anomalies inside another and that were highly similar in nature,…”

We have changed the sentence to “Similar anomalies inside of another, such as (2a,b) in Fig.2.C, were grouped together as one anomaly.” We hope this is more understandable.

*LINE 252: “… an spatial accuracy errors in the gaze.” Should be “… the spatial accuracy errors in the gaze.”

We have changed “an” to “the”.

*Caption figure 3: “… could incorrectly labeled as blinks. “. Something about the construction of this sentence is not right.

You are correct, “be” was missing as well as a comma. We have now changed the sentence to “The particular subject in 3a had a high tracking ratio, though many data samples could be incorrectly labeled as blinks.”

*LINES 329-330: “Since baseline data was the two second fixation cross directly before each stimulus…”. Do you mean: “Since baseline data was sampled during the two seconds the fixation cross was displayed….”?

Yes, and we have rewritten the line based on your suggestion.

*LINES 445-446: Please revise the following sentence: “Whereas bottom-up strategies that student generally employ focus on salient, noticeable images features regardless of relevancy and is less efficient.”

We have changed the sentence to “Whereas bottom-up strategies that student generally

employ is less efficient, as focus is on salient, noticeable images features, regardless of

relevancy.”

*LINES 499-501: Please revise the following sentence: “Experts, renowned for their streamlined processing abilities, are able to selectively allocate their attention to relevant information and is evident in the pupillary response.”

We have condensed the sentence to “Experts are able to selectively allocate their attention to relevant information and is evident in the pupillary response”.

Attachment

Submitted filename: Response to Reviewers.docx

Decision Letter 2

Susana Martinez-Conde

13 May 2020

Pupil diameter differentiates expertise in dental radiography visual search

PONE-D-19-27421R2

Dear Dr. Castner,

We are pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it complies with all outstanding technical requirements.

Within one week, you will receive an e-mail containing information on the amendments required prior to publication. When all required modifications have been addressed, you will receive a formal acceptance letter and your manuscript will proceed to our production department and be scheduled for publication.

Shortly after the formal acceptance letter is sent, an invoice for payment will follow. To ensure an efficient production and billing process, please log into Editorial Manager at https://www.editorialmanager.com/pone/, click the "Update My Information" link at the top of the page, and update your user information. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to enable them to help maximize its impact. If they will be preparing press materials for this manuscript, you must inform our press team as soon as possible and no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

With kind regards,

Susana Martinez-Conde

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Acceptance letter

Susana Martinez-Conde

15 May 2020

PONE-D-19-27421R2

Pupil diameter differentiates expertise in dental radiography visual search

Dear Dr. Castner:

I am pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please notify them about your upcoming paper at this point, to enable them to help maximize its impact. If they will be preparing press materials for this manuscript, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

For any other questions or concerns, please email plosone@plos.org.

Thank you for submitting your work to PLOS ONE.

With kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Prof. Susana Martinez-Conde

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Fig. Pupillary response over course of experiment.

    The average pupillary response from baseline for students (blue bars, 20 images total) and experts (red bars, 15 images total) during the first set of OPTs presented and the second set of OPTs presented. Their is no effect in the pupillary response that could be attributed to fatigue during the experiment.

    (PDF)

    S1 Table. Table of Expert and Student Gaze Counts.

    Shows the gaze hits on each anomaly type for both students and experts. For both levels of expertise, the least difficult and intermediate have the most gaze hits. The following are the ambiguous and the most difficult anomalies. Students had overall more gaze hits than experts; however, this may be attributed to the 90 second viewing time they had in comparison to the 45 second viewing time that the experts had.

    (PDF)

    Attachment

    Submitted filename: Response to Reviewers.docx

    Attachment

    Submitted filename: Response to Reviewers.docx

    Data Availability Statement

    The data and analysis scripts from the presented study are publicly available at: ftp://peg-public:peg-public@messor.informatik.uni-tuebingen.de/peg-public/norac/VisualExpertiseRadiology.zip.


    Articles from PLoS ONE are provided here courtesy of PLOS

    RESOURCES