Abstract
Rapid Automatized Naming (RAN) is strongly related to literacy gains in developing readers, reading disabilities and reading ability in children and adults. Because successful RAN performance depends on the close coordination of a number of abilities, it is unclear what specific skills drive this RAN-reading relationship. The current study used concurrent recordings of young adult participants’ vocalizations and eye movements during the RAN task to assess how individual variation in RAN performance depends on the coordination of visual and vocal processes. Results showed that fast RAN times are facilitated by having the eyes one or more items ahead of the current vocalization, as long as the eyes do not get so far ahead of the voice as to require a regressive eye movement to an earlier item. These data suggest that optimizing RAN performance is a problem of scheduling eye movements and vocalization given memory constraints and the efficiency of encoding and articulatory control. Both RAN completion time (conventionally used to indicate RAN performance) and eye-voice relations predicted some aspects of participants’ eye movements on a separate sentence reading task. However, eye-voice relations predicted additional features of first-pass reading that were not predicted by RAN completion time. This shows that measurement of eye-voice patterns can identify important aspects of individual variation in reading that are not identified by the standard measure of RAN performance. We argue that RAN performance predicts reading ability because both tasks entail challenges of scheduling cognitive and linguistic processes that operate simultaneously on multiple linguistic inputs.
Keywords: Rapid Automatized Naming (RAN), eye tracking, sentence reading, task scheduling
In the Rapid Automatized Naming (RAN) task (Denckla & Rudel, 1974) a participant is presented with a grid of familiar stimuli (drawn from sets of letters, numbers, colors or objects) and must name the stimulus items out loud in order as quickly and accurately as possible. Speed and accuracy on this task are strongly related to future literacy gains by preliterate children (NELP, 2008; Denckla & Rudel, 1974), reading disabilities such as dyslexia (Denckla & Rudel, 1976, Wolf & Bowers, 1999), and reading ability in children, adolescents and adults (Arnell, Joanisse, Klein, Busseri & Tannock, 2009; Powel, Stainthorp, Stuart, Gardwood & Quinlan, 2007; Kuperman & Van Dyke, 2011; Swanson, Trainin, Necochea & Hammill 2003). Successful performance on the RAN task requires visual recognition of individual stimuli, access to phonological codes and rapid articulation. Sustained attention is required to manage perceptual encoding and vocal execution according to available working memory capacity in order to optimize speed while minimizing interference between successive items. Individual variation in RAN performance could depend on any or all of these cognitive task components as well as the ability to coordinate these processes.
One approach to understanding RAN performance is to explore how it is related to performance on tasks that measure its component cognitive processes. Although RAN appears to be more strongly correlated with oral reading compared to silent reading (Georgiou, Parrila, Cui & Papadopoulos, 2013; Moll, Fusseneger, Willbuger & Landerl, 2009), articulation rate in tasks involving well-known sequences (e.g., counting to 10) is not typically associated with reading ability (Di Filippo et al. 2005) or with RAN performance (Cutting & Denckla, 2001). The efficiency of other lower-level perceptual and motor processes does appear to contribute to RAN performance (Wolf & Bowers, 1999). For example, slow RAN performance has been found to relate to poor performance on simple visual processing tasks such as visual same/different judgments (Stainthorp et al. 2010). Although performance on phonological processing tasks is highly predictive of reading ability, phonological ability has a low-to-modest relationship to RAN performance (Swanson et al. 2003) and the RAN-reading relation is at least partially independent of phonological processing (Bowers & Newby-Clark, 2002; Wolf & Bowers, 1999; Wolf, Bowers & Biddle, 2000). Orthographic processing does show some relation to RAN performance (Georgiou, Parrila & Kirby, 2009; Cutting & Denckla, 2001 for discussion), a pattern that is consistent with stronger prediction of reading ability by the symbolic (letter and digit) RAN tasks than the non-symbolic (object and color) RAN tasks (Kuperman & Van Dyke, 2011; Maya, Katzir, Wolf & Poldrack, 2004). Nonetheless, object and color RAN tasks also predict reading ability (Rudel & Denckla, 1974) and all of the RAN tasks correlate highly with each other (Bowers & Swanson, 1991; Arnell et al., 2009). As a result, successful RAN performance has been hypothesized to depend on the ability to form strong visual-verbal links in memory (Kirby, Georgiou, Martinussen, & Parrila, 2010), which in turn may facilitate rapid lexical access (Georgiou et al., 2009). The items in the RAN task are typically arranged in a way that is similar to how the words in a text are arranged in the reader’s language (e.g., left to right for English). This similarity has led to suggestions that in part the RAN is related to reading at least because it provides a sensitive measure of how well the eyes move and attention shifts when processing a sequence of items with this layout (Clarke, Hulme & Snowling, 2005; Kuperman & Van Dyke, 2011). However, the association between RAN performance and reading is equally strong when RAN items are named in reverse order as in the usual order (Scarborough & Domgaard, 1998; Protopapas, Atanki & Georgiou, 2013), a finding that suggests that any contribution of scanning ability to the RAN-reading relation does not depend on the specific pattern of eye movements that has been practiced through reading.
Beyond the task-specific skills required for successful RAN performance, RAN has been considered an indicator of more general cognitive skills. Speed of performance in other complex tasks (e.g., cross out and visual matching tasks) is strongly related to RAN performance (Kail, Hall & Caskey, 1999), though the RAN-reading relationship persists after controlling for processing speed (Cutting & Denckla, 2001). Measures of visual attention, attentional shifting, and inattention show some relation to RAN or are impaired in populations that show poor RAN performance (Waber, Wolff, Forbes & Weiler, 2000; Hari & Renvall, 2001; Pham, Fine, & Semrud-Clikeman, 2011) suggesting that RAN tasks provide a measure of executive control (Denckla & Cutting, 2001). The serial nature of the RAN task is important to its relation to reading as discrete naming tasks do not consistently predict reading skill (Perfetti, Finger & Hogaboam, 1978; Stanovich 1981; Bowers 1995; Georgiou et al., 2013) and serial naming tasks discriminate more robustly than discrete naming tasks between dyslexic and non-dyslexic readers (Jones, Branigan & Kelly, 2009). Moreover, serial naming shows a stronger relation to reading after accounting for variation in performance on isolated naming tasks (Logan, Schatschneider & Wagner, 2011). Serial processing of simultaneously presented stimuli may tap into executive attentional processes involved in the dynamic memory updating and the cognitive suppression of previous responses, as indicated by the relationship between RAN and Rapid Serial Visual Presentation (RSVP) tasks (Arnell et al., 2009). Finally, RAN taps working memory as shown by its association with performance on sentence-span (Georgiou, Das & Hayward, 2008; Swanson & Kim, 2007) and other complex-span tasks (Kuperman & Van Dyke, 2011). In sum, RAN shows relations to performance on a wide variety of tasks, but as reviewed by Kirby et al. (2010), none of these measures can fully explain the RAN-reading relationship, which has been found to persist in studies that have controlled for non-RAN articulation rate, processing speed, phonological short-term memory, phonological awareness, and orthographic processing skill, as well as various measures of attention and memory.
A different approach to understanding RAN performance has been to separately evaluate eye movements and vocalizations as they occur sequentially within the RAN task itself in relation to overall RAN performance. Vocal recordings of RAN can be decomposed into Articulation Time (AT) and Pause Time (PT), with the idea that AT depends on the automaticity of making a verbal response once the item has been recognized, while PT reflects preparation processes related to attention, eye movements and stimulus recognition (Araújo, Inácio, Francisco, Faísca, Petersson & Reis, 2001; Clarke, Hulme, & Snowling, 2005; Neuhaus, Foorman, Francis, and Carlson, 2001). However, there is no consistent evidence on whether AT or PT is the better predictor of reading ability (Clarke et al., 2005; Georgiou, et al., 2009; Georgiou; Papadopoulos, Fella & Parrila, 2012), with Georgiou et al. (2014) demonstrating that the two measures share a large amount of variance and that any unique contribution of PT declines with age, though the relative weight of AT and PT during RAN performance in explaining reading skill in children depends on the orthographic consistency of their language’s writing system (Georgiou, Aro, Liao & Parrila, 2015).
Concurrent recordings of participants’ vocalizations and their eye movements allow for a fine-grained understanding of how RAN performance is shaped by the coordination of visual and vocal processes. Jones and colleagues have examined RAN eye-voice relationships for individuals with dyslexia and non-dyslexic controls using the fixation-speech interval (FSI; Jarvilehto, Nurkkala, Koskela, Holappa & Vierela, 2008; Inhoff, Solomon, Radach & Seymour, 2011) – the time elapsed between when an item is first fixated and when its articulation begins1. In the object RAN task, dyslexic participants showed overall longer FSIs than controls, but visual or semantic similarity between subsequent items led to similar reductions in FSI for both groups (Jones, Branigan, Hatzidaki & Obregon, 2010). In the letter RAN task, phonological-onset similarity between subsequent items was associated with increased FSI for both dyslexic and non-dyslexic participants, suggesting that phonological access was more difficult in the presence of phonologically similar items (Jones, Obregon, Kelly & Branigan; 2008; Jones, Ashby & Branigan, 2013). However, although the FSI as a measure seems analogous to simple stimulus-response time, the sequential constraint of the RAN task requires that articulation of earlier items be completed before articulation of the fixated item can begin. As a result, the FSI is very sensitive to the articulation duration of the preceding item(s) and does not provide a straightforward measure of the time needed to generate an item’s articulatory output after seeing it.
Eye-voice span (EVS) – the number of words that the eyes are ahead of the voice while reading out loud – provides an alternative measure of eye-voice relationships that does not depend on the spoken duration of RAN items. As discussed in Levin and Addis (1979), EVS has long been used in the study of reading and as a tool for assessment and diagnosis of reading deficits. While early studies indicated that EVS was in the range of eight words for mature readers, more recent studies indicate much shorter averages in the range of two words (Inhoff et al., 2011). EVS increases with age and with reading ability (Buswell, 1921; Anderson & Dearborn, 1952; Tinker, 1965; cited by Levin & Addis, 1979), and individuals with dyslexia show similar levels of impairment in EVS as in other measures of reading skill (De Luca, Pontillo, Primativo, Spinelli, Zoccolotti, 2013). Within (non-dyslexic) readers, EVS is reduced when text is more complex (Buswell, 1921; cited by Levin & Addis, 1979) and appears to vary in accordance with phrase or constituent boundaries (Levin & Addis, 1979). As a result, EVS has been hypothesized to reflect the reader’s use of syntactic and semantic information as well as (or in combination with) the temporal interval available for short-term memory (Geyer, 1968; cited by Levin & Addis, 1979). During RAN, eye-voice span can be measured as the number of items that the eyes are fixated ahead of the voice. Pan, Yan, Laubrock, Shu & Kliegl (2013) found that EVS was a significant predictor of performance on the digit RAN task for typically-developing 10-year old children but not for age-matched children with dyslexia. They interpreted this difference in results as indicating that the perceptual span, which determines the region of text surrounding fixation that can be encoded during reading (McConkie & Rayner, 1975; Rayner & Bertera, 1979), is smaller in the dyslexic children than the control children. Similarly, Moll and Jones (2013) have provided evidence that dyslexic readers are less efficient at encoding information beyond the center of fixation than non-dyslexic readers during RAN because they are more sensitive to visual crowding in the parafovea. However, eye movements and reading rate during oral reading are less affected by the availability of parafoveal information than silent reading (Ashby, Yang, Evans & Rayner, 2012). As RAN requires each item to be named out loud, the size or efficiency of readers’ perceptual span is unlikely to fully explain individual differences in RAN performance or the RAN-reading relationship.
Eye-voice span during RAN has also been interpreted as a measure of automaticity, or the extent to which the item-to-sound conversion needed for item naming has moved from a resource-demanding to an overlearned, automatic process (but see Georgiou & Stewart, 2013). Pan et al. (2013) observed that differences in EVS between dyslexic and non-dyslexic children were greater for digit RAN than the rapid naming of dice surfaces, a task that can be assumed to be much less practiced than naming digits. Hogan-Brown, Hoedemaker, Gordon & Losh (2014) observed slower RAN naming and lower EVS for participants with ASD compared to age-matched controls. Siblings of individuals with ASD also exhibited lower EVS than controls, but only on letter and digit RAN, the two conditions that are considered most highly automatized. Siblings of individuals with ASD showed similar performance to controls on color and object RANs, which are assumed to be less-automatized naming tasks.
The present work uses EVS and other eye-voice relations as a source of insight into the ways in which component processes are scheduled or coordinated in time during RAN performance. Analyses of cognitive processes in terms of scheduling began with applications of mathematical scheduling theory to formal analyses of the organization of cognitive processes (e.g., Schweickert, 1978; Schweickert & Boggs, 1984), where it was argued that scheduling models provide a way of characterizing a very important level of cognitive organization that applies across processing domains. The value of analyzing performance on complex tasks in terms of scheduling is widely recognized in models of dual-task performance (Meyer & Kieras, 1997; Pashler, 1994a) and in extension of those models to performance in tasks – like the RAN – that require speeded responses to individual items in a sequence (Pashler, 1994b). Analyzing the RAN in terms of scheduling naturally incorporates the straightforward notion that having the eyes lead the voice allows upcoming RAN items to be perceptually encoded so that their identities are available when processes of articulatory planning and control are ready to use that information. It further leads to a focus on the fact that this simultaneous processing on successive RAN items must be coordinated in time – scheduled – so that the eyes are sufficiently far ahead of the voice to supply the needed information but not so far ahead that the number of encoded-but-yet-to-be-produced items is too large to be effectively held in working memory. It is likely that the extent to which scheduling constrains performance on RAN tasks depends on the efficiency of the mechanisms that must operate on each individual RAN item, such as those for visual identification, name retrieval and articulatory planning/execution, and also on the efficiency of general mechanisms that directly support scheduling, such as working-memory capacity and executive function (Engle, 2002; Kane & Engle, 2003; Miyake & Friedman, 2012). The substantial body of research showing that the RAN-reading relationship survives after controlling for individual variation in the component skills of RAN performance suggests that individual variation in the efficiency of the general mechanisms that support the scheduling of component processes may make an important contribution to the predictive value of the RAN task. By analyzing RAN performance and the RAN-reading relationship in terms of scheduling, this paper aims to provide a framework that can incorporate the specific component processes of the RAN and silent reading as well as address the ways in which these component processes must be coordinated to enable the type of efficient sequential processing that characterizes skilled reading.
During silent reading the eyes provide information about the time course of early orthographic and lexical processes but later processes of lexical integration and comprehension cannot be measured directly, making it difficult to study the ways in which early and later processes of reading and comprehension are coordinated. In contrast, eye movements during the RAN task provide an overt measure of the time course of perceptual encoding of linguistic units, and the acoustic properties of the participants’ speech provide information about the time course of subsequent articulatory processes. Thus, the relationship between the eyes and the voice in the RAN task provides information about the scheduling of earlier and later processes that is not available for silent reading. The current research studies a group of skilled readers (college students) in order to determine whether – and how – eye-voice relations in the RAN predict individual differences in RAN performance and in eye movements during silent reading. EVS provided the initial eye-voice measure of interest in the RAN task, but exploration of the data showed the importance of considering the incidence of regressive saccades in explaining individual differences in RAN completion time and in the efficiency of word recognition during silent reading.
Method
Participants
Fifty undergraduates at the University of North Carolina at Chapel Hill participated in the study in exchange for course credit in Introductory Psychology. Data from two participants were not usable because of problems with their audio recordings, leaving data for 48 participants for the analyses reported below. All participants had normal or corrected-to-normal vision, and none indicated being colorblind. All participants were naïve to the purpose of the study.
Apparatus
An SR EyeLink 1000 (0.25°–0.5° average accuracy) was used to record eye movements from the participants’ dominant eye. All stimuli were presented on a 22″ Samsung LCD monitor with 120 Hz refresh rate and a resolution of 1680 × 1050 at a viewing distance of approximately 22″. In order to accommodate the maximum visual angle of the eye tracker, stimulus presentation was confined to a 12″ × 10″ portion of the screen. During the RAN task, vocal responses were recorded on the PC using a table-mounted microphone with an E-MU 0404 USB amplifier (Creative), and an ASIO sound card. Audio recording began automatically at the beginning of each trial.
RAN task
Stimuli for the Rapid Automatized Naming Task were taken from the Comprehensive Test of Phonological Processing (CTOPP, Wagner, Torgesen & Rashotte, 1999)2. Each trial contained an array of 36 items arranged on a grid for four rows and nine columns. There were two trials (A and B) for each of the four RAN types: colors (blue, red, green, black, brown, yellow); letters (s, t, n, a, k, c); objects (pencil, star, fish, chair, boat, key) and numbers (2, 3, 4, 5, 7, 8). On each trial, the items on the CTOPP are arranged in a pseudorandom order (with the restriction that no item appears in two consecutive positions) that is fixed for all participants. The B trials reverse the order of the A trials. Each RAN type was preceded by a practice trial containing one instance of each of the six items that would appear for that type of RAN, followed by the A and B trial.
The color items were adjusted slightly from the original source to render them more easily distinguishable on the LCD screen. In addition, the size and spacing of the individual items was adjusted to maximize use of the available room on the display. Areas of Interest (AOIs) were assigned to each item by drawing a virtual grid on the array, so that each item’s AOI extended from its center to halfway toward its neighboring items. The full RAN grid extended across 31 × 26 degrees of visual angle and each AOI took up approximately 3.5 degrees of horizontal visual angle. Within each AOI, the sizes of the various stimulus items in degrees of visual angle were as follows: color (3×4), letters (1×1.5), objects (2×2.5) and digits (1×1.5).
Participants were instructed to name all items in the array as quickly and accurately as possible, naming the items in each row from left to right starting with the top row.
Reading task
Participants read 10 warm-up sentences and then read a set of 40 sentences that were 10-to-15 words long. The sentences were taken from a larger pool of sentences used for a different study, and were constructed as follows. One hundred sets of four target words were selected so that each set contained two high frequency and two low frequency words. High and low frequency status was defined within each HF-LF word pair, so that each pair differed by at least one unit of log frequency (SUBTLEXus, Brysbaert & New, 2009). Average log word frequency was 3.24 per 51 million for the HF and 1.81 per 51 million for the LF words. Each four-word set was assigned two different sentence frames, each containing two possible slots for the four words from the set. High and low frequency words were rotated so that each sentence could appear with two HF words, two LF words, or a mix of one HF and one LF word (see Table 1 for an example of a set of sentence frames and target words). The resulting 200 sentences were divided into four lists, so that all sentence frames appeared once in each list, but with different target words across lists. As a result, no target words or sentence frames were repeated within a list. Each list was divided into 5 blocks of 40 sentences each, resulting in a total of twenty counterbalanced blocks. Each participant in the current experiment was presented only one block of 40 sentences, rotating through the 20 available blocks across participants. The sentences appeared in a different random order for each participant.
Table 1.
Sample sentence frames with interchangeable high/low frequency target words.
| 1a. | My thoughtful father/cousin gave me some onions/gourds from his own garden |
| 1b. | Anna forgot her father/cousin did not like onions/gourds until it was too late. |
Note: Target words are italicized here but were presented in regular font during the experiment.
The alternative high and low frequency words at a given target location were matched for number of letters, with target words having a mean length of 6.5 letters and a range of 5 to 9 letters. The stimulus sentences were constructed with the goal that the target words would not be predictable. After the main experiment, predictability was assessed using a cloze task in which a separate group of participants was presented with sentence fragments and asked to ‘continue [each fragment] to create a complete sentence’. Sentence fragments were presented up to but not including the first target word (e.g., “Anna forgot her…”) and in two versions up to but not including the second target word (one version containing the low-frequency first target and one version containing the high-frequency first target, e.g., “Anna forgot her father/cousin did not like…”). These three conditions provided measures of the cloze probability of the first target word and of the second target word for the two conditions in which it could appear. These conditions were counterbalanced across groups of participants so that participants only saw each sentence in one of the three conditions. All participants in the cloze task were native speakers of English selected from the same participant pool as those who participated in the eye-tracking portion of the experiment. Cloze data were obtained from 20 to 30 participants for each stimulus. Predictability for the target words was low (mean of 0.78% and 3.23% respectively for low- and high-frequency words), though this difference was statistically significant, t(398) = 5.63, p < .001. However, these mean levels of predictability are far below those used in experimental studies that have manipulated predictability of target words. For example, across a number of experiments examining how predictability (or constraint) affects eye movements during reading, the mean cloze values were 64%, 86%, 41%, 64% and 78% (Balota, Pollatsek & Rayner, 1985; Rayner & Well, 1996; Rayner, Binder, Ashby & Pollatsek, 2001). The predictability of individual target words in the current stimulus set ranged from zero to 59% for low-frequency words and zero to 64% for high-frequency words, showing that predictability was not avoided for all 400 target locations in the stimulus set (200 sentences with two target locations each). However, subsidiary analyses to be reported below show that the subset of predictable words had very little effect on the reading data.
Sentences were presented on a single line in black 20-point Times New Roman font on a white background. Despite the use of a proportional font there was no difference in mean physical size of the low-frequency and high-frequency target words (mean of 74.3 pixels for high-frequency words and 74.0 pixels for low-frequency words, F < 1). One degree of visual angle spanned approximately 3 letters (34.3 pixels). Participants were instructed to read for comprehension at a natural pace. Twenty-five percent of the sentences were followed by a true/false comprehension question, which participants answered by pressing a button on a handheld console.
Testing procedure
Participants were seated in a dimly-lit, sound-attenuated room. They completed the sentence-reading task and then the RAN task. During the sentence-reading portion of the study, a chin rest and forehead rest were used to minimize head movements. During the RAN, only the forehead rest was used to allow for movement of the jaw during vocal responses. Both the RAN and the reading experiment began with a nine-point calibration procedure. Average spatial calibration error did not differ with (.46 degrees) and without (.45) the chinrest, t(47) = .06, p > .5. Each trial began with a fixation point on the left side of the screen, marking the location where the first word of the sentence would appear, or in the upper left corner, marking the location of the first RAN item. The experimenter initiated each trial as soon as the participant established a steady gaze on the fixation point. During the RAN task, subjects advanced to the next trial by pressing a button on a handheld console after naming all 36 items. During the reading portion of the experiment, participants pressed the same button after reading each sentence. Gaze location was monitored throughout the experimental session and recalibrated as necessary.
Results
Analysis procedures for RAN
Automatic phonetic alignment of the vocal responses was performed using the Penn Phonetics Lab Forced Aligner (Yuan & Liberman, 2008). Subsequently, the onset and offset boundaries output by the automatic alignment procedure were manually edited by two (hypothesis-blind) trained coders based on both the visual waveform and auditory assessment of the response. Errors were identified as any response that deviated from the serial naming of all 36 items, including item-name substitutions, repetitions, skips, self-corrections and any other type of extraneous speech (e.g., ‘eh,’ ‘ehm’ and other fillers).
Vocal responses and associated eye movements for the first item and the last items were excluded from all analyses because of difficulties in interpreting eye-movements in relation to those items. As a result, completion time is defined as the time between the onset of the second item to the offset of the 35th item on each trial. Total duration is the duration of an individual vocal response plus its preceding silence. Silence duration is the duration of the inter-word silence. If there was no silent gap detected in between two subsequent vocalizations, silence duration was coded as zero.
Fixations shorter than 80 ms and within 1 degree of a longer, immediately subsequent fixation were merged with the longer fixation by an automatic procedure in the EyeLink software. Due to track loss no fixations were recorded for a portion of a small number of items (1.02% of all items). These items were excluded from all analyses. Eye-Voice Span (EVS) was defined as the number of items that the eyes were ahead of the voice at the onset of the vocal response. Vocal responses with an EVS of less than 0 (suggesting that the participant was fixating an item that occurred before the currently named item) or more than 5 (suggesting that the participant was looking more than 5 items ahead of the currently pronounced item) were excluded from the analyses (0.43% of all items). Such EVS values deviated substantially from canonical task performance, and were likely due to tracker error and/or erratic, off-task behavior. Regressions to previous items were defined as any fixation on an item after one or more subsequent items had already been fixated. Regression rate is the proportion of fixated RAN items from which there was a regressive eye movement to an earlier item in the sequence. Skips were defined as any occasion on which a RAN item was not fixated at all, or not until after the fixation of a later item. Skipping rate refers to the proportion of RAN items that were skipped. Return saccades (RS) were leftward saccades from near the end of one line toward the beginning of the next line. These long saccades were associated with decreased spatial accuracy; most often resulting in undershoot of the target or short return saccades (SRS). As a result, SRSs were often followed by one or more regressive saccades that were corrections of short returns.
Statistical methods
Analysis of variance was used to assess the statistical significance of differences in the means of these measures across the different types of RAN tasks. Subsequently, linear regression was used to determine whether eye-voice measures could explain individual differences in RAN completion time.
RAN time and accuracy
Table 2 shows basic results for vocal performance in the RAN task. Consistent with previous research (Bos, Zijlstra & Spelberg, 2002; Cronin & Carver, 1998; Denckla & Rudel, 1974; Hogan-Brown et al., 2014; Meyer, Wood, Hart, & Felton, 1998; Misra, et al., 2004), average completion time was longer for the Object RAN than for the Color RAN, F(1,47) = 59.3, p < .001, longer for the Color RAN than for the Letter RAN, F(1,47) = 192.9, p < .001, and longer for the Letter RAN than for the Digit RAN, F(1,47) = 38.2, p < .001. The average number of errors per RAN trial was low for these college students, on average showing less than one error per two trials. Accordingly, error rates will not be analyzed further. However, it should be kept in mind that completion time encompasses time spent on erroneous responses, self-corrections and fillers. As such, completion time will be considered a comprehensive measure of RAN performance. Decomposition of the RAN time showed that both the average durations for correctly spoken items and the average durations of silence preceding correctly spoken items were longer for the Object RAN than for the Color RAN, F(1,47) = 24.4, p < .001, F(1,47) = 21.2, p < .001, longer for the Color RAN than for the Letter RAN, F(1,47) = 363.7, p < .001, F(1,47) = 27.2, p < .001, and longer for the Letter RAN than for the Digit RAN, F(1,47) = 21.3, p < .001, F(1,47) = 22.0, p < .001.
Table 2.
Summary of vocal response measures across RAN types.
| Completion Time (ms) Mean (SD) |
Errors Per Trial Mean (SD) |
Duration of Vocal Response (ms) Mean (SD) |
Duration of Silence if Present (ms) Mean (SD) |
|
|---|---|---|---|---|
| Object | 21866 (3881) | .45 (.552) | 433 (52) | 265 (98) |
| Color | 19093 (3472) | .40 (.447) | 410 (48) | 219 (76) |
| Letter | 13826 (3439) | .32 (.378) | 312 (57) | 165 (77) |
| Digit | 12572 (2996) | .10 (.204) | 294 (57) | 129 (72) |
| Mean | 16839 (3002) | .32 (.245) | 362 (48) | 194 (66) |
Eye movements and Eye-Voice Span (EVS) during RAN
Qualitatively, the pattern of eye movements during the RAN, shown in Table 3, is broadly similar to the pattern found during reading, a result no doubt of the sequential nature of both tasks. Like most words, most RAN items are fixated at least once; however, occasionally some are skipped, though less frequently than words are skipped during reading (Rayner, 1998). There was a non-significant increase in the proportion of items skipped from Object to Color, to Letter and to Digit RAN. In addition, while the eyes generally move forward, they occasionally move backwards, though again less frequently in the RAN than is typically observed during reading (Rayner, 1998). Regression rate was higher for the Object RAN than for the Color RAN, F(1,47) = 5.9, p = .019, did not differ for the Color and Letter RANs, F(1,47) < 1, and was higher for Letter RANs than for Digit RANs, F(1,47) = 7.6, p = .008. Table 3 shows that the proportion of return saccades that landed short of the first item in the row was .58, with no significant differences in the proportion of short returns as a function of RAN type. The proportion of short returns followed by a corrective saccade to the first item in the row was .74. Proportion of short returns that were corrected did not differ between the Object and Color RANs, F(1,37) = 1.2, p > .25, was significantly higher for the Color RAN than the Letter RAN, F(1,37) = 14.9, p < .001, and did not differ significantly between the Letter and Digit RANs, F(1,37) < 13. The finding that corrections of short returns were less common for symbolic (Letter and Digit) RANs than for non-symbolic (Object and Color) RANs suggests that the symbolic RAN items could be recognized more easily without direct fixation than could the non-symbolic RAN items.
Table 3.
Summary of eye movement measures across RAN types. Skipping rate refers to the proportion of items that did not receive a first-pass fixation. Regression rate refers to the proportion of fixated RAN items from which there were regressive eye movements to earlier items in the sequence. Proportion short return saccades refers to the proportion of return saccades that undershot the target RAN item (the first item on the new line). Proportion of short return saccades corrected refers to the proportion of short returns that were corrected by a regressive saccade aimed towards the original target of the return saccade.
| Skipping rate Mean (SD) |
Regression Rate Mean (SD) |
Proportion Short Return Saccades Mean (SD) |
Proportion of Short Returns Corrected Mean (SD) |
EVS Mean (SD) |
|
|---|---|---|---|---|---|
| Object | .047 (.026) | .085 (.058) | .53 (.31) | .83 (.26) | 0.95 (.13) |
| Color | .057 (.037) | .067 (.034) | .58 (.28) | .84 (.22) | 0.94 (.14) |
| Letter | .084 (.059) | .072 (.060) | .65 (.31) | .66 (.37) | 1.01 (.22) |
| Digit | .097 (.059) | .055 (.059) | .58 (.36) | .62 (.38) | 1.23 (.31) |
| Mean | .071 (.034) | .070 (.042) | .58 (.24) | .74 (.25) | 1.03 (.16) |
Figure 1 shows the proportion of articulation onsets on which the eyes led the voice by zero, one, or two-or-more items. For all four RAN types the modal EVS was one item, occurring for 75.9% of the articulations averaged across RAN types but showing lower values of 72.5% and 65.6% for the Letter and Digit RANs respectively. These decreases in frequency of the modal EVS are attributable to a greater incidence of longer EVSs (two-or-more items) for the Letter and Digit RANs. Table 3 shows the mean EVS for all RANs and for each RAN separately. Mean EVS did not differ significantly for the Object and Color RANs, F(1,47) < 1, but was lower for the Color RAN than for the Letter RAN, F(1,47) = 6.4, p = .015, and lower for the Letter RAN than for the Digit RAN, F(1,47) = 53.8, p < .001.
Figure 1.
Average proportion of vocalization onsets with an eye-voice span (EVS) of 0, 1, or 2-or-more across RAN types.
Figure 2 shows eye-voice traces for individual trials in the Letter RAN for three participants. Hypotheses about the ways that eye-voice relations contribute to RAN performance were generated by examining such traces. Comparison of the top and middle panels of Figure 2 illustrates a consistent difference wherein completion time decreased as EVS increased (e.g., the trial from Participant 15 had greater average EVS and shorter completion time than did the trial from Participant 37). This relationship was found previously by Pan et al. (2013) for typically developing 10-year olds but not for those with dyslexia. Further exploration of these data showed that RAN performance was also affected by regression rate - the likelihood that the eyes move backwards to an earlier item; this effect can be seen in the bottom panel of Figure 2 where Participant 29 has a long completion time and frequent regressive saccades. A straightforward interpretation of the pattern shown by Participant 29, and by other similar participants, is that having the eyes ahead of the voice facilitates RAN performance as long as the eyes are not so far ahead as to require regressing to an earlier item in order to avoid confusion. Below we test this interpretation by examining how EVS and the rate of regressive saccades interact in predicting individuals’ RAN completion time and by examining the local impact of regressive saccades on the vocal duration of RAN items.
Figure 2.
Eye-voice traces for individual trials of the Letter RAN for three participants. The position in terms of RAN items (vertical axis) across time (horizontal axis) is shown with green lines for the eyes and blue lines for the voice. The insets provide a blow-up of the pattern for RAN items 10 through 18. EVS (measured in number of RAN items) is indicated by the magnitude of the vertical gap between the green and blue lines. Participant 15 (top panel) has a large average EVS; it can be seen that at the onset of vocalizing an item the eyes are almost always fixating on a RAN item that is one or two ahead. Participant 37 (middle panel) has a smaller average EVS and it is apparent that at times the voice catches up to the eyes, so that at the onset of vocalizing an item the eyes are fixating that same item. Participant 29 (bottom panel) often has a substantial lead of the eyes over the voice but has frequent regressive saccades that bring the eyes back to an earlier item.
An initial series of linear regression models was performed in order to evaluate the extent to which individual variation in average RAN completion time across RAN types could be predicted from EVS and regression rate4. Collapsing across RAN types provides a stable measure of RAN performance, as shown by the high test-retest reliability for participants’ completion times (r = 0.95, R2 = 0.90), EVS (r = 0.88, R2 = 0.78) and regression rate (r = 0.78, R2 = 0.60). As noted above, the pattern of regressive saccades is strongly influenced by the line position of the item because of corrective regressions that occurred following return saccades at the end of the line. This pattern of eye movements related to line returns likely causes the relationship between the eyes and the voice to differ for the central items of a row and for the peripheral items at both ends of a row. Accordingly, statistical analysis of the relationship of RAN performance and reading to RAN eye-voice relations was performed separately for central (columns 3–7) and peripheral items (columns 1–2 and 8–9), with the exception that columns 1 and 2 of the first row and columns 6 through 9 of the final (fourth) row were excluded from all analyses because they exhibited distinctive eye-voice patterns associated with the beginning and the end of the trial.
Table 4 shows the results of linear regression models that predict RAN completion time for the RANs based on EVS (Model 1) and on both EVS and regression rate (Model 2); these models were assessed separately using eye-voice measures from central and peripheral items. For Model 1, the relationship between EVS and completion time fell short of statistical significance for central items but was highly significant for peripheral items. For Model 2, regression rate was a significant predictor of completion time for both central and peripheral items. Further, adding regression rate to the model strengthened the relation between EVS and completion time as compared to Model 1 where EVS was the sole predictor. Prediction was strongest for eye-voice measures taken from peripheral items, where they accounted for 48% of the variance in completion time. The stronger prediction for eye-voice measures taken from peripheral items suggests that this pattern is particularly important when coordinating vocal responses around long return saccades. Participants who were able to keep the eyes consistently ahead of the voice across line breaks achieved faster RAN times than those who did not.
Table 4.
Model fits for RAN completion times as a function of Eye-Voice Span (Model 1) and of Eye-Voice Span and regression rate (Model 2).
| Central Items | Peripheral Items | ||||||
|---|---|---|---|---|---|---|---|
| B | SE | t | B | SE | t | ||
|
| |||||||
| Model 1 | EVS | −4.73 | 2.70 | −1.75 | −6.33 | 1.97 | −3.21** |
|
| |||||||
| R2 | .06 | .18 | |||||
|
| |||||||
| Model 2 | EVS | −5.34 | 2.43 | −2.19* | −6.99 | 1.60 | −4.37*** |
|
| |||||||
| Regr | 49.13 | 14.23 | 3.45** | 17.10 | 3.38 | 5.04*** | |
|
| |||||||
| R2 | .26 | .48 | |||||
p < .05;
p < .01;
p < .001
More generally, the results of these initial linear-regression models support the interpretation that having the eyes lead the voice facilitates RAN performance as long as the eyes do not lead by so much that regressive saccades are required in order to avoid confusion. This interpretation was tested by examining the immediate effects of regressive saccades on eye-voice relations and on the durations of RAN items. As in the determination of EVS (see above), the onset of vocalization of a RAN item is used as the temporal landmark for determination of eye-voice relations. A regressive saccade was deemed to have occurred during a RAN item if it occurred after the onset of vocalization of that item and before the onset of vocalization of the next item. Identifying the concurrent RAN item in this way provides a rough measure of the temporal relation between the regressive saccade and the vocalization sequence. This temporal relation can be refined by dividing regressive saccades into final and non-final regressions. For final regressions, the regressive saccade was the last saccade that occurred during the concurrent RAN item; its landing position determined the fixation location at the onset of the vocalization of the next RAN item. For non-final regressions, the regressive saccade was followed by a progressive saccade before the onset of vocalization of the next RAN item; this means that there was a regression-progression sequence during the concurrent RAN item. Because the vast majority of regressive saccades in the RAN data were followed immediately by a progressive saccade (94.2%), the difference between final and non-final regressions is matter of timing with respect to the spoken sequence of RAN items. The relationship of regression condition (final regression, non-final regression or no regression) to eye-voice measures and to the durations of nearby RAN items was evaluated in order to provide an indication of how regressive eye movements affect RAN performance within a RAN sequence.
The first column of data in Table 5 shows the relation between regression classification and the maximum by which the eyes moved ahead of the voice during the concurrent RAN item (maximum lead). For analyses over all RAN regions a highly significant relation was observed, F(2,86) = 17.3, p < .001, with the mean of the maximum lead being lower for vocalizations that were not concurrent with a regression as compared to the average maximum lead during vocalizations concurrent with either of the two regression classifications, t(43) = 7.45, p < .001, and with the mean of the maximum lead being lower for non-final regressions as compared to final regressions, t(43) = 2.82, p = .0075. The finding that regressions are associated with higher maximum eye-voice leads is consistent with the argument that regressive saccades occur because the eyes have gotten too far ahead of the voice. In addition, there was a highly significant relationship between regression classification and duration of the current RAN item, F(2,86) = 55.0, p < .001, with shorter durations when there was no regression as compared to the average of the two regression classifications, t(43) = 9.46, p < .001, and longer durations for non-final regressions as compared to final regressions, t(43) = 5.49, p < .001. When the peripheral RAN items were considered alone, the same statistical pattern was observed, F(2,80) = 24.2, p < .001, with smaller maximum eye-voice leads when there were no regressions as compared to when there were regressions, t(40) = 6.12, p < .001, and with maximum leads being smaller for non-final as compared to final regressions, t(40) = 4.23, p < .001. Durations for peripheral RAN items varied significantly with regression classification, F(2,80) = 28.2, p < .001, again with shorter durations when there was no regression as compared to the average of the two regression classifications, t(40) = 6.96, p < .001, and longer durations for non-final regressions as compared to final regressions, t(40) = 4.42, p < .001. For central RAN items, the relationship between the mean of the maximum lead and regression pattern was not significant, F(2,50) < 1. The absence of a significant effect of regression classification on EVS for the central items considered alone may be due to the small number of regressive saccades. Nonetheless, durations for central RAN items varied significantly with regression classification, F(2,50) = 23.4, p < .001, again with shorter durations when there was no regression as compared to the average of the two regression classifications, t(25) = 6.59, p < .001, and longer durations for non-final regressions as compared to final regressions, t(25) = 2.62, p = .015.
Table 5.
Relations between regression classification, maximum lead (the maximum number of items the eyes moved ahead of the voice during the concurrent RAN item), duration of the concurrent RAN item (vocalized RAN item during which the regressive saccade occurred), EVS (number of items the eyes were ahead of the voice at the onset of the next RAN item) and vocal duration of the next RAN item for all RAN regions combined and when assessed separately for central and peripheral RAN items. Values in parentheses are standard deviations.
| Regression Type | Measure | ||||
|---|---|---|---|---|---|
|
| |||||
| Maximum Lead | Duration Current RAN Item | EVS | Duration of Next RAN Item | ||
| All RAN Regions | None | 1.89 (.192) | 546 (101) | 0.89 (.193) | 503 (91) |
| Non-Final | 2.03 (.179) | 903 (264) | 0.97 (.192) | 489 (107) | |
| Final | 2.25 (.470) | 701 (230) | 0.45 (.290) | 566 (154) | |
| Peripheral Regions | None | 1.90 (.275) | 567 (118) | 0.90 (.276) | 545 (118) |
| Non-Final | 2.04 (.177) | 900 (319) | 0.99 (.212) | 477 (111) | |
| Final | 2.43 (.621) | 661 (278) | 0.63 (.370) | 554 (167) | |
| Central Regions | None | 1.88 (.173) | 569 (113) | 0.88 (.173) | 518 (92) |
| Non-Final | 1.93 (.360) | 926 (248) | 0.88 (.323) | 511 (131) | |
| Final | 1.92 (.245) | 782 (275) | 0.09 (.168) | 634 (298) | |
The remaining data in Table 5 show the relationships between regression classification, mean EVS and the duration (vocalization plus following silence) of the RAN item following the RAN item being spoken over the interval in which regression classification was assessed (“next RAN item”). Over all RAN regions, mean EVS for the next RAN item varied significantly by regression type, F(2,86) = 99.6, p < .001. Mean EVS was higher when there was no regression as compared to the mean of the two regression classifications, t(43) = 7.73, p < .001, but it is apparent from the pattern of EVS that this difference is entirely due to a substantial reduction in mean EVS when the regression was the final saccade before the next vocalization compared to when it was non-final, t(43) = 10.57, p < .001. The duration of the next RAN item was significantly related to regression type, F(2,86) = 7.54, p = .001, with a non-significant trend for shorter durations when there was no regression than when there was, t(43) = 1.81, p = .08, and with shorter durations when the regression was non-final as compared to when it was final, t(43) = 3.03, p = .004. The patterns of differences and statistical significance for analyses on the peripheral and central RAN items were very similar to those for the overall analyses. Peripheral RAN items showed a significant relation between regression type and mean EVS, F(2,80) = 25.6, p < .001. EVS was higher when there was no regression compared to when there was, t(40) = 2.49, p < .05, an effect that was due to the reduction in EVS when the regressive saccade was final compared to when it was non-final, t(40) = 6.00, p < .001. Duration of the next RAN item was significantly related to regression type, F(2,80) = 6.75, p <.01, with a non-significant trend for longer durations when there was no regression compared to when there was t(40) = 1.75, p = .09, but significantly shorter durations when the regression was non-final compared to when it was final, t(40) = 2.96, p < .01. Central RAN items also showed a significant relation between regression type and EVS, F(2,50) = 119.5, p < .001. EVS was higher when there was no regression compared to when there was, t(25) = 10.97, p < .001, and again this effect was due to the reduction in EVS when the regressive saccade was final compared to when it was non-final, t(25) = 10.92, p <.001. Duration of the next RAN item was significantly related to regression type, F(2,50) = 4.0, p < .05. While there was no overall difference in duration when there was a regression compared to when there was not, t(25) = 1.71, p = .1, in cases where there was a regression, duration of the next RAN item was significantly shorter when the regression was non-final compared to when it was final, t(25) = 2.11, p < .05.
The manner in which different types of regressions are related to the time course given by local measures of RAN performance is readily interpretable in terms of the task demands. For cases of non-final regressions, the regressive saccade is followed by a forward saccade before the onset of vocalization for the next RAN item. These additional saccades are associated with longer durations for the concurrent RAN items as compared to when there is no regression or when there are final regressions. However, the additional forward saccade advances the eyes so that the EVS for the next RAN item is close to the modal value of one (see Figure 1) and very similar to the EVS found when there is no regression at all. The duration of the RAN item following non-final regressions is no longer than for RAN items following cases where there was no regression at all, which can be taken as an indication that the disruption in processing associated with the regression has been resolved by that point and that normal processing has been resumed. The time course is quite different for cases where the regression is the final saccade before the onset of the next RAN item. As discussed, the duration of the current RAN item is longer than when there was no regression, but it is shorter than for cases of non-final regressions where there was an additional forward saccade. However, having a final regressive saccade affects the following RAN item in such a way that the modal EVS (one item) is achieved for less than half of the items. In addition, the duration of the RAN item following final regressions is longer than those following non-final regressions, indicating that the processing disruption associated with the regression has not yet been resolved and is continuing to affect performance.
The substantially better fits that were obtained with both regression rate and EVS (Model 2) as compared to EVS show that regression rate is an important predictor of RAN completion time. Further, inclusion of regression rate strengthened the predictive relationship between EVS and completion time, which suggests that not accounting for regressions suppressed that relationship. The finding (Table 5) that RAN items have longer durations when they coincide with regressive saccades suggests that including regression rates improves model fits through this direct association with longer durations on specific RAN items. While that is likely true to some extent, the low overall regression rate suggests that the contribution of regression rate to model fit might also be based on processes beyond those that are observed locally in the form of increased vocal duration during or shortly after a regressive saccade. To test this possibility, the models using EVS and regression rate to predict total RAN completion time were rerun, but the outcome measure of total RAN completion time was replaced by the average duration (vocalization plus following silence) of RAN items that did not coincide with or immediately follow a regressive saccade. As seen in Table 6, regression rate continued to make substantial contributions to eye-voice models of RAN performance even when the outcome measure did not include RAN items whose durations were directly affected by regressive saccades. This suggests that regressive saccades are both a direct indication of a disruption in smooth processing during the RAN task and also an indirect indication of the individual’s susceptibility to processing disruptions that do not result in this overt behavior.
Table 6.
Model fits for average duration (vocalization plus any following silence) of RAN items for which there were no regressive saccades during the vocalization of that item or the preceding item. Fits are shown for the participants’ Eye-Voice Span (Model 1) and for their Eye-Voice Span and their regression rate (Model 2) as measured when the RAN items associated with regressions were included.
| Central Items | Peripheral Items | ||||||
|---|---|---|---|---|---|---|---|
| B | SE | t | B | SE | t | ||
|
| |||||||
| Model 1 | EVS | −142.36 | 76.77 | −1.85 | −203.63 | 54.59 | −3.73** |
|
| |||||||
| R2 | .07 | .232 | |||||
|
| |||||||
| Model 2 | EVS | −157.43 | 71.33 | −2.21* | −220.45 | 45.89 | −4.80*** |
|
| |||||||
| Regr | 1220.38 | 416.96 | 2.93** | 439.50 | 97.01 | 4.53*** | |
|
| |||||||
| R2 | .22 | .47 | |||||
p < .05;
p < .01;
p < .001
Reading - Analysis procedures and word-recognition results
Fixations shorter than 80 ms and within one degree of a longer, immediately subsequent fixation were merged with the longer fixation by an automatic procedure in the EyeLink software. Reading times for the two target words in each sentence were analyzed according to several standard measures of eye movements during reading. Skipping rate is the proportion of occasions on which a word did not receive a first-pass fixation, meaning that the word was never fixated or that it was only fixated after the eyes had already gone beyond the word. Skipping rate is affected both by oculomotor factors (e.g., saccade overshoot) and ease of processing a word in the parafovea to the right of fixation (Vitu, O’Regan, Inhoff & Topolsky, 1995; Rayner, Sereno & Raney, 1996; Gordon, Plummer & Choi, 2013; Choi & Gordon, 2013; 2014). Gaze duration is the sum of all first-pass fixations on a target word, starting when the eyes first enter the word’s region and ending when the eyes exit the region either to the left or the right; it is generally regarded as the most appropriate eye-movement measure of word recognition (Inhoff, 1984; Morris, 1994; Rayner, 1998). First-pass regression rate is the proportion of occasions on which the first-pass fixations on a target word are followed by a regressive eye movement to an earlier region of text; it is thought to reflect difficulty in integrating the target word into the evolving sentence meaning and in some instances correction for an earlier mis-targeted saccade (Frazier & Rayner, 1982; Rayner, Sereno, Morris, Schmauder & Clifton, 1989).
Mean accuracy on the post-sentence comprehension questions was 95% (sd = 6.0); no participants’ data were excluded from analyses due to poor question-answering performance. Table 8 shows eye tracking during reading results for the target words as a function of word frequency (high vs. low). Skipping rate was greater when the target word was high frequency as compared to low frequency, F1(1,47) = 8.9, p < .01, F2(1,199) = 9.6, p < .01. Gaze duration was shorter for high- than for low-frequency words, F1(1,47) = 27.3, p < .001, F2(1,196) = 25.6, p < .001. First-pass regression rate was lower for high- than for low-frequency words, F1(1,47) = 5.5, p < .05, F2(1,196) = 7.3, p < .016.
Table 8.
Summary of eye-movement measures on high and low frequency target words during sentence reading.
| Skipping Rate | Gaze Duration | FP Regression Rate | |
|---|---|---|---|
| High-Frequency Words | .154 | 250 | .149 |
| Low-Frequency Words | .119 | 272 | .183 |
RAN performance and word recognition during reading
Table 9 presents regression analyses that show how different facets of RAN performance predict skipping rate, gaze duration and first-pass regression rate during reading and the effects of word frequency on those measures. For each reading measure, models were evaluated using total RAN completion time (the standard RAN measure) and for the two eye-voice models (Model 1: EVS only, Model 2: EVS and regression rate) applied separately to the central and peripheral RAN items.
Table 9.
Model fits for reading measures as a function of total RAN completion time (Total Time) and separately as a function of Eye-Voice Span (EVS, Model 1) and and Eye-Voice Span and regression rate (EVS and Regr, Model 2).
| Prop. Words Fixated | Central Items | Peripheral Items | |||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|
| B | SE | t | B | SE | t | B | SE | t | |||
|
|
|
||||||||||
| Total Time | .01 | .01 | 1.38 | Model 1 | EVS | −.17 | .09 | −1.96 | −.09 | .07 | −1.29 |
|
|
|
||||||||||
| R2 | .04 | R2 | .08 | .04 | |||||||
|
|
|||||||||||
| Model 2 | EVS | −.18 | .09 | −2.03* | −.10 | .07 | −1.38 | ||||
|
| |||||||||||
| Regr | .50 | .47 | 1.07 | .19 | .14 | 1.39 | |||||
|
| |||||||||||
| R2 | .10 | .08 | |||||||||
| Prop. Words Fixated – Frequency Effect | Central Items | Peripheral Items | |||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|
| B | SE | t | B | SE | t | B | SE | t | |||
|
|
|
||||||||||
| Total Time | .00 | .00 | −.09 | Model 1 | EVS | .01 | .08 | .13 | −.01 | .06 | −.20 |
|
|
|
||||||||||
| R2 | .00 | R2 | .00 | .00 | |||||||
|
|
|||||||||||
| Model 2 | EVS | .01 | .08 | .12 | −.01 | .06 | −.15 | ||||
|
| |||||||||||
| Regr | .01 | .42 | .03 | −.10 | .12 | −.81 | |||||
|
| |||||||||||
| R2 | .00 | .02 | |||||||||
| Gaze Duration | Central Items | Peripheral Items | |||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|
| B | SE | t | B | SE | t | B | SE | t | |||
|
|
|
||||||||||
| Total Time | 4.47 | 1.90 | 2.35* | Model 1 | EVS | −92.91 | 35.54 | −2.61* | −49.68 | 28.86 | −1.72 |
|
|
|
||||||||||
| R2 | .11 | R2 | .13 | .06 | |||||||
|
|
|||||||||||
| Model 2 | EVS | −97.47 | 34.11 | −2.86** | −52.01 | 28.67 | −1.81 | ||||
|
| |||||||||||
| Regr | 416.59 | 184.40 | 2.26* | 77.51 | 58.26 | 1.33 | |||||
|
| |||||||||||
| R2 | .22 | .10 | |||||||||
| Gaze Duration – Frequency Effect | Central Items | Peripheral Items | |||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|
| B | SE | t | B | SE | t | B | SE | t | |||
|
|
|
||||||||||
| Total Time | 1.16 | 1.41 | .82 | Model 1 | EVS | −74.25 | 24.49 | −3.03** | −51.27 | 19.56 | −2.62* |
|
|
|
||||||||||
| R2 | .01 | R2 | .17 | .13 | |||||||
|
|
|||||||||||
| Model 2 | EVS | −75.80 | 24.50 | −3.09** | −50.85 | 19.79 | −2.57* | ||||
|
| |||||||||||
| Regr | 141.41 | 132.44 | 1.07 | −13.77 | 40.21 | −.34 | |||||
|
| |||||||||||
| R2 | .19 | .13 | |||||||||
| FP Regression Rate | Central Items | Peripheral Items | |||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|
| B | SE | t | B | SE | t | B | SE | t | |||
|
|
|
||||||||||
| Total Time | .01 | .00 | 3.65** | Model 1 | EVS | −.16 | .08 | −2.14* | −.14 | .06 | −2.42 |
|
|
|
||||||||||
| R2 | .23 | R2 | .09 | .11 | |||||||
|
|
|||||||||||
| Model 2 | EVS | −.17 | .07 | −2.43* | −.15 | .06 | −2.69* | ||||
|
| |||||||||||
| Regr | 1.02 | .39 | 2.62* | .28 | .11 | 2.44* | |||||
|
| |||||||||||
| R2 | .21 | .22 | |||||||||
| FP Regression Rate – Frequency Effect | Central Items | Peripheral Items | |||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|
| B | SE | t | B | SE | t | B | SE | t | |||
|
|
|
||||||||||
| Total Time | −.00 | .01 | −.21 | Model 1 | EVS | .00 | .09 | .01 | .02 | .07 | .33 |
|
|
|
||||||||||
| R2 | .01 | R2 | .00 | .00 | |||||||
|
|
|||||||||||
| Model 2 | EVS | −.00 | .10 | −.01 | .02 | .07 | .24 | ||||
|
| |||||||||||
| Regr | .19 | .56 | .35 | .18 | .16 | 1.15 | |||||
|
| |||||||||||
| R2 | .00 | .03 | |||||||||
p < .05;
p < .01;
p < .001
Readers’ average skipping rate was not predicted by RAN time, a null effect that is consistent with the results of a large scale study conducted around the same time as the current study and using similar materials, with the difference that the RAN task was presented on paper and completion time was measured with a stopwatch (546 readers, Gordon et al. in prep). Of the eye-voice models, only Model 2 applied to the central RAN items showed a significant relation to skipping rate, with higher EVS associated with more skipping. The effect of word frequency on skipping rate was not significantly related to any of the models of RAN performance.
Readers’ average gaze duration declined significantly with their RAN time, a relationship that was also observed in Gordon et al. (in prep) and Kuperman and Van Dyke (2011). Eye-voice Models 1 and 2 both showed significant prediction of gaze duration when applied to central RAN items. Of note, Model 2 (including both EVS and number of regressions) accounted for 22% of the variance in gaze duration as compared to the 11% accounted for by the standard measure of RAN completion time. The effect of word frequency on gaze duration was not related significantly to RAN completion time, again consistent with Gordon et al. (in prep). However, all of the eye-voice models showed significant relations such that a greater EVS was associated with a smaller effect of frequency on gaze duration. This relationship was strongest for models fit to the central items, with Models 1 and 2 accounting for 17% and 19% of the variance respectively.
Readers’ average first-pass regression rate increased significantly with their RAN time, once again showing the same results as in Gordon et al. (in prep). All of the eye-voice models also showed significant relations to first-pass regression rate: greater EVS was associated with lower rates of first-pass regressions and greater RAN regression rate was associated with higher first-pass regression rates in reading. The effect of frequency on first-pass regression rate was not predicted significantly by any of the measures of RAN performance. The models of average first-pass regression rate show three notable features: First, first-pass regression rate was related to the standard measure of RAN completion time as strongly, or more strongly, than to the eye-voice models. Second, first-pass regression rate was predicted equally well by eye-voice models based on central and peripheral RAN items. Finally, within the eye-voice models the statistical relationship of EVS and number of RAN regressions to first-pass regressions appeared equally strong.
Discussion
The scheduling (coordination in time) of eye movements with naming was investigated in relation to individual variation in RAN performance. Across all RAN types, the modal EVS was one item, with symbolic RANs showing greater average EVS than non-symbolic RANs. Across individuals, as well as RAN types, greater average EVS was associated with faster RAN performance and this relationship was strengthened when the number of regressive eye movements was included in the statistical models. Together EVS and regression rate accounted for a quarter (central items) to almost half (peripheral items) of the individual variation in RAN completion times indicating that good coordination of eye-voice relations is particularly important across line breaks. This pattern of results suggests that fast RAN performance is facilitated by having the eyes far enough ahead of the voice that upcoming items can be encoded for vocalization in quick succession, but that progressive eye movements are scheduled in such a way that they do not get so far ahead as to cause confusion and require regressive eye movements. Furthermore, regression rates were predictive of RAN speed on items that were not directly affected by regressive saccades, suggesting that an individual’s tendency to make regressive eye movements serves as an indicator of their susceptibility to processing disruptions even if such disruptions do not always lead to overt changes in behavior. RAN was related to performance on a separate sentence reading task, such that participants with faster RAN completion times showed shorter gaze durations and lower first-pass regression rates during silent reading. Although RAN completion time is conventionally used as the primary indicator of RAN performance, the eye-voice models better accounted for the variation in eye movements during reading. The full eye-voice model accounted for more variance in gaze duration than RAN completion time, and EVS (but not completion time or regression rate) was significantly related to the effect of word frequency, so that participants with larger average EVS showed smaller effects of frequency. RAN completion time and the eye-voice models accounted for equal proportions of variance in regression rate, but the eye-voice model revealed that this relationship results from both EVS and regression rate during the RAN.
Eye-voice relations in the RAN
Figure 3 shows how the relationship in time between looking at a RAN item and saying it determines the degree to which successive RAN items are processed in parallel. These schematic timelines adopt a representation (simultaneous activity on separate items) that has been used extensively in models of how component processes are scheduled during dual-task performance (e.g., Meyer & Kieras, 1997; Pashler, 1994a) and in extensions of those models to serial performance (e.g., Pashler, 1994b). Within each panel time goes from left to right and progress through a sequence of RAN items goes from top to bottom. EVS is shown by the dashed vertical lines; it is the number of items by which the eyes lead the voice at the onset of vocalization of a RAN item (one for the sequence in the top and two for the sequence in the bottom). Comparison of the top and bottom panels of Figure 3 shows that as EVS increases so does the number of RAN items that are processed simultaneously. Thus, while the RAN is a serial task at the level of goals (the items must be said in the correct order), efficient performance is achieved by processing more than one item at a time so that relevant information about downstream items is available without delay when needed by the articulatory system. However, as can be seen in Figure 3, the efficiency of processing downstream items brings with it a requirement that the advance information about items and the order of those items must be accurately stored in working memory until it is used by the articulatory system. This model of RAN performance is consistent with the Protopapas et al. (2013) proposal that developmental changes in RAN performance are driven primarily by an increased ability to process RAN items in a cascaded manner, such that multiple subsequent RAN items are processed simultaneously as they are passed through a ‘pipeline’ of different stages of processing. As pointed out by Protopapas et al., such parallel processing of successive items benefits from the automatization of processes concerned with individual items (visual processing, articulation) but also requires adequate executive control to schedule and monitor the processing of multiple items in parallel processing cascades.
Figure 3.
Schematic timeline showing eye-voice relations across items during the RAN, as well as the memory requirements that are imposed when the eyes leave an item before articulation of the item begins. The top panel illustrates an EVS of one while the bottom panel illustrates an EVS of two. EVS is given by the difference in RAN items of the eyes and voice at the onset of vocalizing an item and is shown by the dashed vertical lines. An EVS of one (as in the top panel) occurs when at the onset of vocalizing an item (the left edge of the voice rectangle) the eyes are already fixating the next item; this relation between the eyes and the voice requires that a single RAN item be stored in memory (as indicated by the thought cloud). An EVS of two (as in the bottom panel) occurs when the eyes are already fixating two items ahead at the onset of vocalization; this relation between the eyes and the voice requires that two items must be stored in memory.
Of course, the memory representations shown in Figure 3 are not fully fleshed out. It is possible that processing each item requires multiple steps (e.g., perceptual encoding, maintenance and articulatory encoding) and therefore that the processing of each item requires storage of multiple representations rather than the single representation shown in Figure 3. Thus, by showing a one-to-one relationship between EVS and the number of items stored in memory Figure 3 may underestimate the rate at which the demand for memory storage increases with EVS. Certainly, the memory loads of one or two items shown in Figure 3 would not of themselves be expected to tax the working-memory capacity of college students (Cowan, 2001; Luck & Vogel, 1997). Furthermore, individual differences in working-memory capacity may affect scheduling efficiency during RAN performance because of the task’s simultaneous demands on memory storage and processes of attention control or executive function (Engle, 2002; Kane & Engle, 2003; Miyake & Friedman, 2012), a possibility that awaits investigation.
Regression models that only included EVS had some ability to predict RAN completion time, but both overall model fit and the strength of the relationship between EVS and RAN completion time increased when the rate of regressive eye movements was included in the model. As discussed previously, a plausible explanation of this pattern is that regressive eye movements are an indication that the eyes got too far ahead of the voice, with the consequence that a regression was needed for the purpose of confirming or correcting the items stored in memory. Support for this view comes from the finding that on average the extent to which the eyes led the voice was greater before regressive saccades than progressive saccades (see Table 5). On this account, performance was best for those participants who were able to keep their eyes sufficiently far ahead that downstream items could be encoded and made available for articulatory processing but not so far ahead that they lost track of intermediate items or item order. However, including regression rate with EVS improved the fit of eye-voice models even for the prediction of individual variation in RAN times that excluded RAN items whose durations were directly affected by regressive saccades (see Table 6). As noted above, this suggests that the rate of regressive saccades is an observable indication of an individual’s susceptibility to disruptions in smooth sequential processing even on occasions where those disruptions do not result in a regressive saccade.
The manner in which eye-voice patterns and RAN completion time varied across RAN types is consistent with individual-differences evidence about how eye-voice relations contribute to RAN performance. As is typically found, RAN completion time decreased from Object to Color to Letter to Digit RAN (Bos, Zijlstra & Spelberg, 2002; Cronin & Carver, 1998; Denckla & Rudel, 1974; Hogan-Brown et al., 2014; Meyer, Wood, Hart, & Felton, 1998; Misra, et al., 2004)). Average EVS was significantly lower for the Object and Color RANs than for the Letter RAN, which was lower than for the Digit RAN. Average regression rate was higher for the Object than for the Color and Letter RANs, which were higher than for the Digit RAN. These differences between symbolic and non-symbolic RANs suggests that the processing stream from encoding to articulation functions more smoothly and is less susceptible to inter-item interference for the Letter and Digit RANs than for the Object and Color RANs (see Figure 1). A stronger association between items and their names for the symbolic compared to the non-symbolic RANs is one possible contributor to this difference though it is likely that other factors (e.g., visual complexity and inherent phonetic length of RAN items) also contribute. The greater EVS for the Digit as compared to the Letter RAN may occur because college students would typically be expected to encode letters during the recognition of words (where letter names play little or no role) while digits are more likely to be encoded or spoken aloud individually. Thus, the steps of perceptual encoding, name retrieval and articulatory planning may work more efficiently for digits than for letters, with visually perceived digits rapidly converted into articulatory codes with fewer opportunities for memory interference than occur for letters or for non-symbolic visual stimuli. This account is consistent with the idea that good RAN performance occurs when naming is automatic (Hogan-Brown et al., 2014; Norton & Wolf, 2012; Pan et al., 2013), but needs to be evaluated in tasks that independently control factors other than the strength of the item-name association.
Eye-voice relations in the RAN for the college students tested in this study showed variation in a small range around a very consistent core pattern (i.e., the EVS of one item that was found for 75.9% of RAN items across all four types of RAN task). The mean magnitude of the EVS is quite similar to those of Pan et al. (2013), who report that their groups’ means for EVS in the digit RAN were .81 for dyslexic ten-year olds (sd = .25) and 1.1 for control ten-year olds (sd = .27). For the adults in the current study mean EVS in the digit RAN was 1.23 (sd = .31). In the Pan et al. study the correlation between RAN completion time and EVS was not significant for the dyslexic 10-year olds, but for the control 10-year olds these two measures were negatively correlated (i.e., faster RAN time associated with larger EVS), as was the case for the adults in the current study. The small differences in EVS between the groups in these two studies show an orderly relation to diagnostic status (dyslexic vs. control) and education level (fifth graders vs. college students). Pan and colleagues interpret their results as indicating that the visual attention span is smaller in the dyslexic children than the control children. The current finding that the strength of the association between EVS and completion time increased when number of regressive eye movements was added to the model (Table 4) indicates that regression rate had a suppressive effect on the relationship between EVS and completion time. Therefore, absence of a relationship between EVS and completion time for Pan et al.’s dyslexic group could be due to suppression of the relationship between EVS and completion time by frequent disruptions of parallel processing of the sort indicated by regressive saccades. The extent to which eye-voice relations in other populations (e.g., very young children) are similar to those found here remains an open question.
RAN performance and skilled reading
The results of this study show that variation in RAN completion time predicted variation in eye-movement measures of lexical processing by skilled readers (see also Gordon et al. in prep). For measures of early word-recognition processes predictions from the eye-voice model of RAN performance were better than predictions based on RAN completion time alone. Predictions for measures of early word recognition were stronger using the eye-voice models for central rather than peripheral RAN items, an outcome that is unsurprising given that the reading study was designed to provide information about recognition of words that were not near the edges of a line. For eye-movement measures of later lexical processing predictions from the eye-voice model and RAN completion time were similar.
With respect to early word-recognition processes, both RAN completion time and eye-voice relations for central RAN items were significantly related to mean gaze duration, though the eye-voice relations accounted for a greater share of variance (22% vs. 11%). Eye-voice relations for central RAN items were significantly related to first-pass skipping and to the effect of word frequency on gaze duration, but RAN completion time was not significantly related to either measure. This pattern is consistent with the findings of the larger study (Gordon et al. in prep) where RAN completion time was significantly related to mean gaze duration but not to the effect of frequency on gaze duration or to skipping rate.
The finding that eye-voice relations in the RAN were a better predictor than RAN completion time for these first-pass measures of early word recognition suggests that eye-voice relations have the potential to provide new insight into the cognitive processes that allow for efficient word recognition during reading. It is possible that this efficiency is found in the actual operation of word-recognition mechanisms, but it seems more likely that it is due to the efficiency with which a reader coordinates early word-recognition processes with rapid first-pass eye movements (e.g., the L1 familiarity check in the EZ Reader model (Reichle, Rayner & Pollatsek, 2003; Pollatsek, Reichl & Rayner, 2006) which is affected by word frequency and determines gaze duration) and to the efficiency with which information gained by those early word-recognition processes is used at subsequent levels of comprehension. Our interpretation of the RAN-reading relationship in adults is consistent with Protopapas et al.’s (2013) proposal that the relationship between RAN and reading across development may be explained by changes in developing readers’ ability to process subsequent items in parallel. Reading entails processing of input at many different levels of abstraction ranging from the processing of visual features, through different levels of word recognition and the integration of meaning across sentences. The manner in which the characteristics of text influence eye movements during reading is consistent with the idea that processing of words occurs faster at lower levels, such as gauging whether a word is familiar, than at higher levels, such as integrating a word’s meaning with an emerging discourse representation (Gordon et al. 2013; Lowder, Choi & Gordon, 2013; Rayner, 1998). For reading to be efficient, these processes must be scheduled so that the ones that are early and quick are completed by the time that their output is needed, but that these early levels do not get so far ahead of the slower processes that their connection to the larger goal of comprehension is lost with the result that reading becomes mindless (McVay & Kane, 2012; Rayner & Fischer, 1986).
With respect to later lexical processing, slightly over 20% of the individual variation in first-pass regression rate during reading was accounted for by RAN completion time, a relationship that was also found in the larger study (Gordon et al. in prep). The additional finding that the two components of the eye-voice model, EVS and regressions rate, predict first-pass regression rate during reading offers insight into why this RAN-reading relationship is observed. The finding that increased rate of regressions in the RAN predicts increased rate of first-pass regressions during reading is consistent with findings showing individual consistency in eye-movement characteristics across tasks (Henderson & Luke, 2014). That is, participants who are more likely to make regressive eye movements in the RAN are more likely to make them during reading. While this finding provides some support for the idea that the RAN-reading relation occurs because both tasks tap the same ability to control eye movements (Clark et al. 2005; Kuperman & Van Dyke, 2011; see Protopapas et al. 2013 for discussion), the finding that increased EVS in the RAN predicts decreased rate of first-pass regressions during reading does not involve a straightforward correspondence of eye movements across the two tasks.
EVS reflects a process in which the eyes are advanced through a sequence so that perceptual encoding can be initiated for upcoming stimuli while processing of earlier items is ongoing. RAN and other reading-aloud tasks allow measurement of both eye and vocal position, with the distance between them giving the span over which processing is taking place. In silent reading the timing of perceptual encoding can be measured reasonably directly by eye position but far less information is available about the timing of processing at higher levels such as selection of word meaning and the integration of meanings within and across sentences. On this interpretation, both RAN performance and silent reading require scheduling of processes that occur simultaneously on a sequence of items and the RAN task is a useful predictor of reading because it indicates the efficiency with which the individual achieves this type of scheduling.
Conclusion
Performance on the RAN task has been shown to be strongly related to reading skill, the acquisition of literacy, and reading disorders. Yet the reasons for its success have been elusive. The analyses of eye-voice relations reported here show that successful performance in the RAN task requires that rapid eye movements and slower articulation be scheduled (coordinated in time) so that the eyes are sufficiently ahead of the voice to allow preparation for upcoming RAN items but not so far ahead as to strain memory for the correct order of the items. It is argued that silent reading also involves coordination of processes that occur at different rates and that this common scheduling challenge is an important factor in the predictive success of the RAN task.
Table 7.
Stimulus characteristics and reading measures after exclusion of items using progressively more stringent criteria. The value in parentheses after the mean Cloze proportion is the proportion of items excluded by the criterion indicated for that row. The r values in parentheses after each reading measure are the correlation between participants’ performance when no items were excluded and when items were excluded using the predictability criterion shown for that row (*p < .05, **p < .01 and ***p< .001 for frequency effect as tested by participants; †p < .05, ††p < .01 and ††† p < .001 for frequency effect as tested by items).
| Cloze Exclusion | WF | Cloze (proportion of items excluded) | Skipping Rate | Gaze Duration | FP Regression Rate |
|---|---|---|---|---|---|
| ≥0.25 | HF | .0318 (.022) | **.154 (r = .998) | ***250(r = .994) | **.122 (r = .939) |
| LF | .0052 (.005) | ††.119 (r = .999) | †††272 (r = .999) | ††.161 (r = .972) | |
| ≥0.10 | HF | .0105 (.101) | **.156 (r = .990) | ***249 (r = .989) | **.123 (r = .919) |
| LF | .0033 (.019) | ††.118 (r = .996) | †††273 (r = .998) | †.162 (r = .972) | |
| > 0.00 | HF | .0000 (.324) | *.147 (r = .922) | ***251 (r = .954) | **.123 (r = .847) |
| LF | .0000 (.115) | †.114 (r = .984) | ††273 (r = .990) | †.165 (r = .969) |
Acknowledgments
This research was support by R01 HD060440-06A2 to PCG from NICHD. We thank Yi Ting Huang, Van Liceralde, Matt Lowder and Giulia Pancani for helpful comments on an earlier version of this manuscript, Shri Narayanan and members of the SAIL LAB at USC for discussion of this project, and members of the Language Cognition and Brain Lab at UNC Chapel Hill for much discussion and help.
Footnotes
In their articles Jones and colleagues call this measure eye-voice span. We use the term fixation-speech interval in order to avoid confusion with a different, well-established meaning for eye-voice span which we discuss below.
This version of the RAN has been characterized along with the Denckla and Rudel (1976) five-by-ten version as using a traditional format for the RAN grid (Compton, Olson, DeFries and Pennington, 2002). Performance on an alternative version of the RAN grid with only five items per row is highly correlated (r = .85) but may account for more variation in word recognition skill in some populations (Compton et al. 2002).
When collapsed across RAN types all 48 participants had non-zero rates of short return saccades but only 38 participants had non-zero rates of short return saccades for every RAN type. The proportion of short returns that are corrected cannot be calculated for conditions where there are no short returns, and for this reason Table 3 shows the mean results for proportion of short returns that are corrected for those 38 participants who had who had non-zero rates of short returns for all RAN types, and this accounts for the reduced degrees of freedom reported for tests between these RAN types on this measure. Performing these tests in a pairwise fashion increases the number of participants included in each analysis but does not change the pattern of results: Object vs. Color (means of .84 and .83), t(43) = .26, p = .80; Color vs. Letter (means of .85 and .64), t(43) = 4.22, p < .001; and Letter vs. Number (means of .66 and .62), t(39) = .72, p = .47.
Our initial explorations of individual differences also examined the different types of RAN tasks separately. Those exploratory analyses showed some tendency for stronger relations between the eye-voice measures and RAN completion time for the symbolic (Letter and Digit) RANs than for the non-symbolic (Object and Color) RANs. However, this tendency was not consistent across measures and the results did not provide a compelling indication about whether the observed patterns reflected true differences between the RANs or whether they simply reflected variation due to sampling error given the quantity of data collected for each participant on each RAN type. Accordingly, our modeling of individual differences uses participants’ results averaged across the four RAN types.
The degrees of freedom are reduced for in statistical tests for differences as a function of the two different types of regressions because only 44 of 48 participants showed both types of regressions. Fewer participants showed both types of regressions when analyses were restricted to peripheral items and fewer still showed both types of regressions for central items where the overall rate of regressions was substantially lower.
The results change little when the analyses were restricted to the less predictable words as shown in the table below where the eye movement results are presented after progressively more stringent criteria are used to exclude the more predictable words (Cloze ≥ 0.25, ≥ 0.10 and > 0). The effects of frequency remain significant both by participants and items under all three exclusion criteria. Further exclusion by the various criteria did not substantially alter the measures at the level of individual participants, as can be seen from the correlations between participants’ results with the full set of items and when items were excluded by the various criteria. These high correlations indicate that individual differences in these measures are robust across different criteria for excluding items due to predictability.
References
- Anderson IH, Dearborn WF. The psychology of teaching reading. New York, NY: The Ronald Press; 1952. [Google Scholar]
- Araújo S, Inácio F, Francisco A, Faísca L, Petersson KM, Reis A. Component processes subserving rapid automatized naming in dyslexic and Non-dyslexic readers. Dyslexia. 2011;17(3):242–255. doi: 10.1002/dys.433. [DOI] [PubMed] [Google Scholar]
- Arnell KM, Joanisse MF, Klein RM, Busseri MA, Tannock R. Decomposing the relation between rapid automatized naming (RAN) and reading ability. Canadian Journal of Experimental Psychology/Revue Canadienne De Psychologie Expérimentale. 2009;63(3):173. doi: 10.1037/a0015721. [DOI] [PubMed] [Google Scholar]
- Bowers PG. Tracing symbol naming speed’s unique contributions to reading disabilities over time. Reading and Writing. 1995;7(2):189–216. [Google Scholar]
- Bowers PG, Newby-Clark E. The role of naming speed within a model of reading acquisition. Reading and Writing. 2002;15(1–2):109–126. [Google Scholar]
- Bowers PG, Swanson LB. Naming speed deficits in reading disability: Multiple measures of a singular process. Journal of Experimental Child Psychology. 1991;51(2):195–219. doi: 10.1016/0022-0965(91)90032-n. [DOI] [PubMed] [Google Scholar]
- Brysbaert M, New B. Moving beyond Kučera and Francis: A critical evaluation of current word frequency norms and the introduction of a new and improved word frequency measure for American-English. Behavior Research Methods. 2009;41(4):977–990. doi: 10.3758/BRM.41.4.977. [DOI] [PubMed] [Google Scholar]
- Buswell GT. The relationship between eye-perception and voice-response in reading. Journal of Educational Psychology. 1921;12(4):217. [Google Scholar]
- Butsch RL. Eye movements and the eye-hand span in typewriting. Journal of Educational Psychology. 1932;23(2):104. [Google Scholar]
- Choi W, Gordon PC. Coordination of word recognition and oculomotor control during reading: The role of implicit lexical decisions. Journal of Experimental Psychology: Human Perception and Performance. 2013;39(4):1032. doi: 10.1037/a0030432. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Choi W, Gordon PC. Word skipping during sentence reading: Effects of lexicality on parafoveal processing. Attention, Perception, & Psychophysics. 2014;76(1):201–213. doi: 10.3758/s13414-013-0494-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Clarke P, Hulme C, Snowling M. Individual differences in RAN and reading: A response timing analysis. Journal of Research in Reading. 2005;28(2):73–86. [Google Scholar]
- Compton DL, Olson RK, DeFries JC, Pennington BF. Comparing the relationships among two difference versions of alphanumeric rapid automatized naming and word leel reading skills. Scientific Studies of Reading. 2002;6(4):343–368. [Google Scholar]
- Cowan N. The magical number 4 in short-term memory: A reconsideration of mental storage capacity. Behavioral and Brain Sciences. 2001;24:87–185. doi: 10.1017/s0140525x01003922. [DOI] [PubMed] [Google Scholar]
- Cronin V, Carver P. Phonological sensitivity, rapid naming, and beginning reading. Applied Psycholinguistics. 1998;19(03):447–461. [Google Scholar]
- Cutting LE, Denckla MB. The relationship of rapid serial naming and word reading in normally developing readers: An exploratory model. Reading and Writing. 2001;14(7–8):673–705. [Google Scholar]
- De Luca M, Pontillo M, Primativo S, Spinelli D, Zoccolotti P. The eye-voice lead during oral reading in developmental dyslexia. Frontiers in Human Neuroscience. 2013;7:696. doi: 10.3389/fnhum.2013.00696. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Denckla MB, Cutting LE. History and significance of rapid automatized naming. Annals of Dyslexia. 1999;49(1):29–42. [Google Scholar]
- Denckla MB, Rudel R. Rapid “automatized” naming of pictured objects, colors, letters and numbers by normal children. Cortex. 1974;10(2):186–202. doi: 10.1016/s0010-9452(74)80009-2. [DOI] [PubMed] [Google Scholar]
- Di Filippo G, Brizzolara D, Chilosi A, De Luca M, Judica A, Pecini C, et al. Rapid naming, not cancellation speed or articulation rate, predicts reading in an orthographically regular language (Italian) Child Neuropsychology. 2005;11(4):349–361. doi: 10.1080/09297040490916947. [DOI] [PubMed] [Google Scholar]
- Engle RW. Working memory capacity as executive attention. Current Directions in Psychological Science. 2002;11(1):19–23. [Google Scholar]
- Frazier L, Rayner K. Making and correcting errors during sentence comprehension: Eye movements in the analysis of structurally ambiguous sentences. Cognitive Psychology. 1982;14(2):178–210. [Google Scholar]
- Georgiou GK, Aro M, Liao CH, Parrila R. The contribution of RAN pause time and articulation time to reading across languages: Evidence from a more representative sample of children. Scientific Studies of Reading. 2015;19(2):135–144. [Google Scholar]
- Georgiou GK, Das J, Hayward DV. Comparing the contribution of two tests of working memory to reading in relation to phonological awareness and rapid naming speed. Journal of Research in Reading. 2008;31(3):302–318. [Google Scholar]
- Georgiou GK, Papadopoulos TC, Fella A, Parrila R. Rapid naming speed components and reading development in a consistent orthography. Journal of Experimental Child Psychology. 2012;112(1):1–17. doi: 10.1016/j.jecp.2011.11.006. [DOI] [PubMed] [Google Scholar]
- Georgiou GK, Papadopoulos TC, Kaizer EL. Different RAN components relate to reading at different points in time. Reading and Writing. 2014;27:1–16. [Google Scholar]
- Georgiou GK, Parrila R, Cui Y, Papadopoulos TC. Why is rapid automatized naming related to reading? Journal of Experimental Child Psychology. 2013;115(1):218–225. doi: 10.1016/j.jecp.2012.10.015. [DOI] [PubMed] [Google Scholar]
- Georgiou GK, Parrila R, Kirby JR. RAN components and reading development from grade 3 to grade 5: What underlies their relationship? Scientific Studies of Reading. 2009;13(6):508–534. [Google Scholar]
- Georgiou GK, Stewart B. Is rapid automatized naming automatic? Preschool & Primary Education. 2013;1(1):67–81. [Google Scholar]
- Geyer JJ. Perceptual systems in reading: The prediction of a temporal eye-voice span constant. In: Smith HK, editor. Perception and reading. Newark, Delaware: International Reading Association; 1968. [Google Scholar]
- Gordon PC, Moore M, Choi W, Hoedemaker RS, Lowder M. Individual differences in reading: Separable effects of knowledge and processing skill. under Review. [DOI] [PubMed] [Google Scholar]
- Gordon PC, Plummer P, Choi W. See before you jump: Full recognition of parafoveal words precedes skips during reading. Journal of Experimental Psychology: Learning, Memory, and Cognition. 2013;39(2):633. doi: 10.1037/a0028881. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hari R, Renvall H. Impaired processing of rapid stimulus sequences in dyslexia. Trends in Cognitive Sciences. 2001;5(12):525–532. doi: 10.1016/s1364-6613(00)01801-5. [DOI] [PubMed] [Google Scholar]
- Henderson JM, Luke SG. Stable individual differences in saccadic eye movements during reading, pseudoreading, scene viewing, and scene search. Journal of Experimental Psychology: Human Perception and Performance. 2014;40:1390–1400. doi: 10.1037/a0036330. [DOI] [PubMed] [Google Scholar]
- Hogan-Brown AL, Hoedemaker RS, Gordon PC, Losh MC. Eye-voice span during rapid automatized naming: Evidence of reduced automaticity in individuals with autism spectrum disorder and their siblings. Journal of Neurodevelopmental Disorders. 2014;6:33. doi: 10.1186/1866-1955-6-33. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Inhoff AW. Two stages of word processing during eye fixations in the reading of prose. Journal of Verbal Learning and Verbal Behavior. 1984;23(5):612–624. [Google Scholar]
- Inhoff AW, Solomon M, Radach R, Seymour BA. Temporal dynamics of the eye–voice span and eye movement control during oral reading. Journal of Cognitive Psychology. 2011;23(5):543–558. [Google Scholar]
- Järvilehto T, Nurkkala V, Koskela K, Holappa E, Vierelä H. Reading as anticipatory formation of meaning: Eye-movement characteristics and fixation-speech intervals when articulating different types of text. Journal of Transfigural Mathematics. 2008;1(1):72. [Google Scholar]
- Jones MW, Ashby J, Branigan HP. Dyslexia and fluency: Parafoveal and foveal influences on rapid automatized naming. Journal of Experimental Psychology: Human Perception and Performance. 2013;39(2):554. doi: 10.1037/a0029710. [DOI] [PubMed] [Google Scholar]
- Jones MW, Branigan HP, Hatzidaki A, Obregón M. Is the ‘naming’ deficit in dyslexia a misnomer? Cognition. 2010;116(1):56–70. doi: 10.1016/j.cognition.2010.03.015. [DOI] [PubMed] [Google Scholar]
- Jones MW, Obregón M, Louise Kelly M, Branigan HP. Elucidating the component processes involved in dyslexic and non-dyslexic reading fluency: An eye-tracking study. Cognition. 2008;109(3):389–407. doi: 10.1016/j.cognition.2008.10.005. [DOI] [PubMed] [Google Scholar]
- Jones MW, Obregón M, Louise Kelly M, Branigan HP. Elucidating the component processes involved in dyslexic and non-dyslexic reading fluency: An eye-tracking study. Cognition. 2008;109(3):389–407. doi: 10.1016/j.cognition.2008.10.005. [DOI] [PubMed] [Google Scholar]
- Kail R, Hall LK, Caskey BJ. Processing speed, exposure to print, and naming speed. Applied Psycholinguistics. 1999;20(02):303–314. [Google Scholar]
- Kane MJ, Engle RW. Working-memory capacity and the control of attention: The contributions of goal neglect, response competition, and task set to Stroop interference. Journal of Experimental Psychology: General. 2003;132(1):47. doi: 10.1037/0096-3445.132.1.47. [DOI] [PubMed] [Google Scholar]
- Kirby JR, Georgiou GK, Martinussen R, Parrila R. Naming speed and reading: From prediction to instruction. Reading Research Quarterly. 2010;45(3):341–362. [Google Scholar]
- Kuperman V, Van Dyke JA. Effects of individual differences in verbal skills on eye-movement patterns during sentence reading. Journal of Memory and Language. 2011;65(1):42–73. doi: 10.1016/j.jml.2011.03.002. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Levin H, Addis AB. The eye-voice span. Cambridge, MA: MIT Press; 1979. [Google Scholar]
- Levin H, Turner A. Sentence structure and the eye-voice span. In: Levin H, Gibson EJ, Gibson JJ, editors. The Analysis of Reading Skill. 1968. pp. 196–220. [Google Scholar]
- Liberman M, Yuan J. Speaker identification on the SCOTUS corpus. Proceedings of Acoustics 2008 [Google Scholar]
- Logan JA, Schatschneider C, Wagner RK. Rapid serial naming and reading ability: The role of lexical access. Reading and Writing. 2011;24(1):1–25. doi: 10.1007/s11145-009-9199-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lonigan C, Schatschneider C, Westberg L The National Early Literacy Panel; National Institute for Literacy, editor. Developing Early Literacy: Report of the National Early Literacy Panel. 2008. Identification of children’s skills and abilities linked to later outcomes in reading, writing, and spelling; pp. 55–106. [Google Scholar]
- Lowder MW, Choi W, Gordon PC. Word recognition during reading: The interaction between lexical repetition and frequency. Memory & Cognition. 2013;41:738–751. doi: 10.3758/s13421-012-0288-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Luck SJ, Vogel EK. The capacity of visual working memory for features and conjunctions. Nature. 1997;390:279–281. doi: 10.1038/36846. [DOI] [PubMed] [Google Scholar]
- McConkie GW, Rayner K. The span of the effective stimulus during a fixation in reading. Perception & Psychophysics. 1975;17(6):578–586. [Google Scholar]
- McVay JC, Kane MJ. Why does working memory capacity predict variation in reading comprehension? On the influence of mind wandering and executive attention. Journal of Experimental Psychology: General. 2012;141(2):302. doi: 10.1037/a0025250. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Meyer DE, Kieras DE. A computational theory of executive cognitive processes and multiple-task performance: Part 1. Basic Mechanisms. Psychological Review. 1997;104:3–65. doi: 10.1037/0033-295x.104.1.3. [DOI] [PubMed] [Google Scholar]
- Meyer MS, Wood FB, Hart LA, Felton RH. Longitudinal course of rapid naming in disabled and nondisabled readers. Annals of Dyslexia. 1998;48(1):89–114. [Google Scholar]
- Miyake A, Friedman NP. The nature and organization of individual differences in executive functions: Four general conclusions. Current Directions in Psychological Science. 2012;21(1):8–14. doi: 10.1177/0963721411429458. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Misra M, Katzir T, Wolf M, Poldrack RA. Neural systems for rapid automatized naming in skilled readers: Unraveling the RAN-reading relationship. Scientific Studies of Reading. 2004;8(3):241–256. [Google Scholar]
- Moll K, Fussenegger B, Willburger E, Landerl K. RAN is not a measure of orthographic processing. evidence from the asymmetric german orthography. Scientific Studies of Reading. 2009;13(1):1–25. [Google Scholar]
- Morris RK. Lexical and message-level sentence context effects on fixation times in reading. Journal of Experimental Psychology: Learning, Memory, and Cognition. 1994;20(1):92. doi: 10.1037//0278-7393.20.1.92. [DOI] [PubMed] [Google Scholar]
- Neuhaus G, Foorman BR, Francis DJ, Carlson CD. Measures of information processing in rapid automatized naming (RAN) and their relation to reading. Journal of Experimental Child Psychology. 2001;78(4):359–373. doi: 10.1006/jecp.2000.2576. [DOI] [PubMed] [Google Scholar]
- Norton ES, Wolf M. Rapid Automatized Naming (RAN) and reading fluency: Implications for understanding and treatment of reading disabilities. The Annual Review of Psychology. 2012;63:427–452. doi: 10.1146/annurev-psych-120710-100431. [DOI] [PubMed] [Google Scholar]
- Pan J, Yan M, Laubrock J, Shu H, Kliegl R. Eye–voice span during rapid automatized naming of digits and dice in chinese normal and dyslexic children. Developmental Science. 2013;16(6):967–979. doi: 10.1111/desc.12075. [DOI] [PubMed] [Google Scholar]
- Pashler H. Dual-task interference in simple tasks: Data and theory. Psychological Bulletin. 1994a;116(2):220. doi: 10.1037/0033-2909.116.2.220. [DOI] [PubMed] [Google Scholar]
- Pashler H. Overlapping mental operations in serial performance with preview. The Quarterly Journal of Experimental Psychology. 1994b;47(1):161–191. doi: 10.1080/14640749408401148. [DOI] [PubMed] [Google Scholar]
- Pollatsek A, Reichle ED, Rayner K. Tests of the EZ reader model: Exploring the interface between cognition and eye-movement control. Cognitive Psychology. 2006;52(1):1–56. doi: 10.1016/j.cogpsych.2005.06.001. [DOI] [PubMed] [Google Scholar]
- Perfetti CA, Finger E, Hogaboam TW. Sources of vocalization latency differences between skilled and less skilled young readers. Journal of Educational Psychology. 1978;70(5):730. [Google Scholar]
- Pham AV, Fine JG, Semrud-Clikeman M. The influence of inattention and rapid automatized naming on reading performance. Archives of Clinical Neuropsychology : The Official Journal of the National Academy of Neuropsychologists. 2011;26(3):214–224. doi: 10.1093/arclin/acr014. [DOI] [PubMed] [Google Scholar]
- Powell D, Stainthorp R, Stuart M, Garwood H, Quinlan P. An experimental comparison between rival theories of rapid automatized naming performance and its relationship to reading. Journal of Experimental Child Psychology. 2007;98(1):46–68. doi: 10.1016/j.jecp.2007.04.003. [DOI] [PubMed] [Google Scholar]
- Protopapas A, Altani A, Georgiou GK. Development of serial processing in reading and rapid naming. Journal of Experimental Child Psychology. 2013;116(4):914–929. doi: 10.1016/j.jecp.2013.08.004. [DOI] [PubMed] [Google Scholar]
- Rayner K. Eye movements in reading and information processing: 20 years of research. Psychological Bulletin. 1998;124(3):372. doi: 10.1037/0033-2909.124.3.372. [DOI] [PubMed] [Google Scholar]
- Rayner K, Bertera JH. Reading without a fovea. Science (New York, NY) 1979;206(4417):468–469. doi: 10.1126/science.504987. [DOI] [PubMed] [Google Scholar]
- Rayner K, Fischer MH. Mindless reading revisited: Eye movements during reading and scanning are different. Perception & Psychophysics. 1996;58(5):734–747. doi: 10.3758/bf03213106. [DOI] [PubMed] [Google Scholar]
- Rayner K, Sereno SC, Morris RK, Schmauder AR, Clifton C., Jr Eye movements and on-line language comprehension processes. Language and Cognitive Processes. 1989;4(3–4):SI21–SI49. [Google Scholar]
- Rayner K, Sereno SC, Raney GE. Eye movement control in reading: A comparison of two types of models. Journal of Experimental Psychology: Human Perception and Performance. 1996;22(5):1188. doi: 10.1037//0096-1523.22.5.1188. [DOI] [PubMed] [Google Scholar]
- Reichle ED, Rayner K, Pollatsek A. The EZ reader model of eye-movement control in reading: Comparisons to other models. Behavioral and Brain Sciences. 2003;26(4):445–476. doi: 10.1017/s0140525x03000104. [DOI] [PubMed] [Google Scholar]
- Schweickert R. A critical path generalization of the additive factor method: Analysis of a stroop task. Journal of Mathematical Psychology. 1978;18(2):105–139. [Google Scholar]
- Schweickert R, Boggs GJ. Models of central capacity and concurrency. Journal of Mathematical Psychology. 1984;28(3):223–281. [Google Scholar]
- Stainthorp R, Stuart M, Powell D, Quinlan P, Garwood H. Visual processing deficits in children with slow RAN performance. Scientific Studies of Reading. 2010;14(3):266–292. [Google Scholar]
- Stanovich KE. Relationships between word decoding speed, general name-retrieval ability, and reading progress in first-grade children. Journal of Educational Psychology. 1981;73(6):809. [Google Scholar]
- Swanson HL, Trainin G, Necoechea DM, Hammill DD. Rapid naming, phonological awareness, and reading: A meta-analysis of the correlation evidence. Review of Educational Research. 2003;73(4):407–440. [Google Scholar]
- Swanson L, Kim K. Working memory, short-term memory, and naming speed as predictors of children’s mathematical performance. Intelligence. 2007;35(2):151–168. [Google Scholar]
- Tinker MA. Bases for effective reading. Minneapolis, MN: University of Minnesota Press; 1965. [Google Scholar]
- Van den Bos Kees P, Zijlstra BJ, Lutje Spelberg HC. Life-span data on continuous-naming speeds of numbers, letters, colors, and pictured objects, and word-reading speed. Scientific Studies of Reading. 2002;6(1):25–49. [Google Scholar]
- Vitu F, O’Regan JK, Inhoff AW, Topolski R. Mindless reading: Eye-movement characteristics are similar in scanning letter strings and reading texts. Perception & Psychophysics. 1995;57(3):352–364. doi: 10.3758/bf03213060. [DOI] [PubMed] [Google Scholar]
- Waber DP, Wolff PH, Forbes PW, Weiler MD. Rapid automatized naming in children referred for evaluation of heterogeneous learning problems: How specific are naming speed deficits to reading disability? Child Neuropsychology. 2000;6(4):251–261. doi: 10.1076/chin.6.4.251.3137. [DOI] [PubMed] [Google Scholar]
- Wagner RK, Torgesen JK, Rashotte CA. Comprehensive test of phonological processing: CTOPP. ASHA; 1999. [Google Scholar]
- Wolf M, Bowers PG. The double-deficit hypothesis for the developmental dyslexias. Journal of Educational Psychology. 1999;91(3):415. [Google Scholar]
- Wolf M, Bowers PG, Biddle K. Naming-speed processes, timing, and reading: A conceptual review. Journal of Learning Disabilities. 2000;33(4):387–407. doi: 10.1177/002221940003300409. [DOI] [PubMed] [Google Scholar]
- Yan M, Pan J, Laubrock J, Kliegl R, Shu H. Parafoveal processing efficiency in rapid automatized naming: A comparison between Chinese normal and dyslexic children. Journal of Experimental Child Psychology. 2013;115(3):579–589. doi: 10.1016/j.jecp.2013.01.007. [DOI] [PMC free article] [PubMed] [Google Scholar]



