Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2013 Apr 15.
Published in final edited form as: Vision Res. 2012 Mar 1;59:64–71. doi: 10.1016/j.visres.2012.02.007

Integration time for the perception of depth from motion parallax

Mark Nawrot 1, Keith Stroyan 2
PMCID: PMC3349336  NIHMSID: NIHMS362080  PMID: 22406543

Abstract

The perception of depth from relative motion is believed to be a slow process that “builds-up” over a period of observation. However, in the case of motion parallax, the potential accuracy of the depth estimate suffers as the observer translates during the viewing period. Our recent quantitative model for the perception of depth from motion parallax proposes that relative object depth (d) can be determined from retinal image motion (dθ/dt), pursuit eye movement (dα/dt), and fixation distance (f) by the formula: d/f ≈ dθ/dα. Given the model’s dynamics, it is important to know the integration time required by the visual system to recover dα and dθ, and then estimate d. Knowing the minimum integration time reveals the incumbent error in this process. A depth-phase discrimination task was used to determine the time necessary to perceive depth-sign from motion parallax. Observers remained stationary and viewed a briefly translating random-dot motion parallax stimulus. Stimulus duration varied between trials. Fixation on the translating stimulus was monitored and enforced with an eye-tracker. The study found that relative depth discrimination can be performed with presentations as brief as 16.6 msec, with only two stimulus frames providing both retinal image motion and the stimulus window motion for pursuit (mean range = 16.6–33.2 msec). This was found for conditions in which, prior to stimulus presentation, the eye was engaged in ongoing pursuit or the eye was stationary. A large high-contrast masking stimulus disrupted depth-discrimination for stimulus presentations less than 70-75 msec in both pursuit and stationary conditions. This interval might be linked to ocular-following response eye-movement latencies. We conclude that neural mechanisms serving depth from motion parallax generate a depth estimate much more quickly than previously believed. We propose that additional sluggishness might be due to the visual system’s attempt to determine the maximum dθ/dα ratio for a selection of points on a complicated stimulus.

Keywords: depth perception, motion parallax, pursuit eye movements, binocular stereopsis, motion masking

1. Introduction

Quick and successful locomotion relies on a moving observer’s ability to accurately determine the relative positions of objects in the environment. Motion parallax, the shift in the relative position of these objects created by the observer’s own movement, is a crucial visual cue for this task. The perception of depth from motion parallax relies upon the changing visual scene from the moving observer’s viewpoint. As the observer moves, fixation is maintained on one particular object for some interval while objects nearer than this fixation point appear to move in one direction and more distant objects appear to move in the opposite direction. This relative movement of objects at different distances from the observer’s point of fixation is the proximal retinal cue for motion parallax. An important goal for vision science is to determine how the visual system uses this pattern of retinal image motion to recreate an accurate perception of relative depth for objects in the scene.

The study of depth from motion is complicated by the fact that there are different geometric mechanisms of retinal motion depending on the direction of observer translation and movement of the head and eyes. Motion parallax typically refers to lateral observer translation with the head fixed, observing a rigid scene. This is the case where the combination of smooth eye pursuit and retinal image motion is important (Nawrot & Joyce, 2006). In contrast, the extensive study in “optic flow” usually involves some forward observer translation that gives rise to a different geometry. For example, when translation is straight toward the fixate the resulting retinal flow is expansion from a point, but when observer translation is purely lateral, perpendicular to the line of sight, there is no focus of expansion in the retinal image motion. Recognition of this fundamental dichotomy goes as far back as von Kreis, writing in (von Helmholtz 1910, vol. III, Note 4, p. 371), “… The changes of which he (von Helmholtz) speaks are such as the observer would notice if he advanced forward without changing the attitude of his head or his eyes especially. In reality the phenomena are complicated by the fact that, supposing our attention is attracted, not by some object moving along with us, but by stationary external objects, we are invariably in the habit of keeping the eyes fastened for a brief space on some definite point, by turning them so as to counteract the effect of the forward motion of the body….” von Kries concludes with, “Now these apparent motions are just as useful as those described by Helmholtz for forming estimates of distance; and the probability is that both of them generally contribute to the result in some way, although it would be hard to say exactly how.” While the visual system may process the visual cues from these different geometries in similar ways, by restricting our study to lateral observer translation and excluding the effects of forward translation, we are beginning to understand the neural mechanisms serving depth from of motion parallax and unravel von Kreis’ dichotomy.

Echoing von Kreis’ dichotomy, our mathematical model is restricted to lateral observer translation whereas several classic papers (such as Koenderink and van Doorn (1976) and Longuet-Higgins and Prazdny (1980)) are set up to cover general motion. However, many specific results in that literature in are undefined in the case of lateral translation. For example, the relative depth formula (Eq. 2.13) of Longuet-Higgins and Prazdny (1980) is undefined for lateral translation because W=0 and W appears in a denominator. (W represents forward translation of the observer. Their vanishing point is “at infinity” in the case of lateral translation and the formulas that use it are undefined. The results on forward motion are interesting, but that is the other half of von Kreis’ dichotomy.) We could cast our results in their general framework (see Nawrot & Stroyan, 2009, p.1972 and Stroyan & Nawrot, 2011, section 5), but our model simplifies the depth formula in our analysis and uses a direct representation of the retinal motion and smooth pursuit cues studied in our experiments.

More importantly, even in the lateral translation case, where the formulas do not depend on division by W (e.g., Nakayama & Loomis, 1974), the perception of relative depth in these analyses depends on perception of observer translational velocity. Indeed, much of the classic theory on depth from motion (Nakayama & Loomis, 1974; Longuet-Higgins & Prazdny, 1980; Hanes et al., 2008) specifically relies on observer translation speed in relative depth formulas. However, the source of this information is not well specified, and it is clearly not a direct visual cue. Moreover, the difficulty is that the same visual flow is also proposed to be a source of information about observer translation (e.g., “self-motion”, Lappe et al., 1999). The connection between this “observer velocity” approach and our current approach is explained in (Nawrot & Stroyan, 2009, Section 2.2.).

In another classic approach (e.g., Gordon, 1965; Koenderink & van Doorn, 1976; 1987; Perrone & Stone, 1994; Fermüller & Aloimonos, 1997; Domini & Caudek, 1999) the geometry of optic flow is used to derive relative depth strictly from retinal information. This approach does not invoke observer translation velocity or an extra-retinal signal. All of the information is retinal. In a sense, Freeman and Fowler (2000) combine both approaches in experiments on perception of slant with and without pursuit. In the case of observation with pursuit, they write that, “…the visual system must rely on extra-retinal information about translation speed…” but do not give the mechanism for this perception nor use that quantity directly in their theoretical analysis, but rather seem to assume it replaces the fixed-eye retinal translation flow in their model. However, in their conclusion they write, “We have assumed throughout that motion-based slant perception involves separate estimates of relative motion … and (observer) translation speed.” So, while they demonstrated the significance of an extra-retinal pursuit signal in the perception of slant, it is unclear what specific role pursuit might serve. Subsequent research (Nawrot 2003; Naji & Freeman 2004; Nawrot & Joyce 2006) demonstrates the necessary role of pursuit in unambiguous perception of depth sign in motion parallax.

In contrast to these “observer velocity” and “retinal information” approaches, the novel feature of our theoretical approach (Nawrot & Stroyan, 2009; Stroyan & Nawrot, 2011) is the elimination of a need to assume the observer directly perceives his translation velocity, but instead uses the ratio of one retinal (motion) and one non-retinal (pursuit) cue as the primary mathematical quantity of the analytical model. There is growing neurophysiological and psychophysical evidence that this combination of retinal image motion and a pursuit signal is necessary (Nadler et al., 2008) and sufficient (Nawrot & Stroyan, 2009; Nadler et al., 2009; Kim et al., 2011; Moeeny & Cumming, 2011) for the case of lateral observer translation. Our basic model for the perception of unambiguous scaled depth from motion parallax is as follows: Consider the schematic below (Figure 1) describing one simple example condition that produces motion parallax. In this example an observer translates rightward while maintaining gaze on the fixate (F) at viewing distance (f) along the line perpendicular to the line of translation. We assume the eye crosses this line at time t=0. In (Figure 1) a distractor (D) is shown more distant (d) than the fixate also along this line. As the observer translates, fixation on point F produces a change in gaze angle α at the rate /dt (or change in a small time dt). It appears that corresponds to an internal signal from the pursuit eye movement system (Nawrot & Joyce, 2006; Nadler et al., 2009). Observer translation causes the retinal position of point D, described by the angle θ, to change at the rate /dt (or change in a small time dt). This corresponds to the retinal image motion of D, moving in relation to the fixation point.

Figure 1.

Figure 1

Depiction of the geometry of motion parallax. The left panel depicts an eye viewing a point of fixation (F) at a viewing distance (f). The goal is to estimate the relative depth (d) of the distractor point (D). The right panel shows the two angles created by rightward observer translation at the rate dx/dt. The magnitude of pursuit corresponds the change in gaze angle α at the rate /dt. The retinal motion of point D corresponds to the change in angle θ at the rate /dt.

The precise geometric relation between the fixation distance (f), retinal image motion (), and pursuit eye movement () and relative depth (d) is given by the Motion/Pursuit Law (Stroyan & Nawrot, 2011):

d=dθdα11dθdαf (1)

at time t=0, but this more complicated expression is approximately equal to the simple approximation of the Motion/Pursuit Ratio:

dfdθdαordθdαf (2)

at time t=0 if the relative depth is small. For example, if the ratio is 0.1, then the difference between the M/PLaw’s calculation and the M/PRatio’s estimate of d is 10%.

Formulas (1) and (2) each account for both the direction and the magnitude of perceived relative depth of any distractor point. To expand on the case shown in Figure 1, objects nearer and farther than the fixation point generate opposite directions of retinal image motion (± ) and therefore generate estimates of d that are in opposite directions. A change in the direction of observer translation reverses the sign of both and , preserving the ratio, and giving constancy to the estimate of d. Changes in the point of fixation result in changes in f and . Therefore, the model can generate different relative depth estimates (d) for each object while using the same singular values for the fixation distance ( f) and pursuit () that reflects a specific point of fixation.

This time-varying nature of the motion/pursuit ratio is also important to the applicability of our model to perception. It is important to know the integration time (t) required by the human visual system to recover dθ/dt, process dα/dt, and then generate an estimate of relative depth (d). While these mathematical models use instantaneous retinal image and pursuit velocities, the human visual system requires a temporal interval to compute the velocities. But how long does that integration interval need to be?

The perception of depth from motion has been considered sluggish, and required to “build-up” over a period of observation (e.g., 750 msec, Eby, 1992; 800 – 1000 msec, Hildreth et al., 1990; 600 – 1000 msec, Andersen & Bradley, 1998; 700- 1000 msec, Treue et al. 1991), especially for the perception of 3-D structure-from-motion such as when viewing a rotating kinetic depth figure (Wallach & O’Connell, 1953). If this processing sluggishness is true of motion parallax, then the potential accuracy of the depth estimates suffers as the translation of the observer changes the spatial relationship between the points, as viewed from the observer’s vantage point. For instance, two equidistant objects are no longer equidistant as an observer moves in any direction other than directly between them. Moreover, even the viewing distance to the point of fixation changes as the observer translates. This effect of observer motion would not be an issue if the depth-from-motion process were instantaneous. But as this visual processing mechanism takes some interval of time, some amount of error or depth confusion becomes inherent to this depth perception process. Knowing these temporal limits will provide us an indication of the inherent depth perception limitations in motion parallax.

The current study seeks to determine the temporal parameters for the judgment of depth and motion of a motion parallax stimulus. Given the unusual challenges of studying temporal characteristics of perceptual processes, a high-contrast pattern-masking paradigm was used. That is, measuring the stimulus presentation duration required to perform a perceptual task is a poor measure of the underlying mechanism’s temporal characteristics. This is because perceptual processing may continue after stimulus presentation has ceased. Similarly, reaction time measures are inappropriate because they rely on the characteristics of the motor system. For this reason the study of temporal characteristics of perceptual processes has given rise to a variety of masking paradigms (Breitmeyer, 1984). The rationale is that the subsequent presentation of a high-contrast pattern (the mask) can interfere with the perceptual processing of the immediately preceding stimulus. However, the interference effect of the masking stimulus has a limited time frame. If the perceptual processing is complete, the masking stimulus has no effect. The goal with a masking paradigm is to determine the stimulus presentation duration, or stimulus onset asynchrony (SOA), in which observer psychophysical performance is not affected by the mask. This is does not reveal the exact the duration of a visual process, but it does indicate a stage of completion to which the process is unaffected by the mask. That is, a masking paradigm allows the comparison of SOAs between different conditions and allows the conclusion that different processes have different temporal aspects. In the current study, the SOAs for masked and unmasked stimuli are compared for the conditions in which the observer must discriminate depth from motion, discriminate the direction of stimulus motion driving pursuit, and discriminate the direction of local relative motion within the stimulus. If the motion and depth conditions have similar SOAs, one might conclude that the temporal limitation on the depth perception process is found in motion processing. However, if motion and depth have different SOAs, one might conclude that depth requires additional processing, which we propose includes the integration of an internal pursuit signal to disambiguate the perception of depth from motion.

2. Methods

2.1. Apparatus

Stimuli were presented on a 21” Sony Trinitron CRT with a 120 hz refresh-rate (8.3 ms per frame) controlled by a Macintosh computer. Observers viewed from 57 cm with their head held in a headrest. Observers initiated trials and entered responses using the computer keyboard.

Eye position was monitored throughout the experiment with a video-based eye tracking system (Applied Science Laboratories, Bedford, MA ) running under the control of a dedicated Windows-based computer. The eye tracking system sent monocular gaze position coordinates to the experimental computer through an analog connection from the eye tracking system. The experimental computer received the analog eye position coordinates using a 16-bit multifunction I/O board (National Instruments, Austin, TX). Prior to each block of trials, observers completed a 9-position calibration routine for the eye tracker control software. This was followed by a two-position gaze calibration for the experimental computer system.

2.2. Stimuli

In the current experiment stimuli translated across the monitor face while the observer remained stationary and pursued a fixation point at the center of the stimulus. This simplifies the motion parallax paradigm by requiring only pursuit eye movements, which correspond to dα. In contrast, a motion parallax paradigm relying on observer head translation invokes a combination of both pursuit and translational vestibular ocular response eye movements. Only the pursuit signal is believed to be the extra-retinal signal involved in the perception of unambiguous depth from motion parallax (Nawrot & Joyce, 2006).

The design of motion parallax stimuli has been detailed previously (Nawrot & Joyce, 2006). In the current study motion parallax stimuli comprised 4000 2 min2 dots within a stimulus window size of 6.6 × 6.6 deg. The stimulus window translated horizontally with a fixed velocity of 4.0 deg/sec. The stimulus depicted 1.5 cycles of depth corrugation, with 0.75 cycles above and below the center point. To generate these depth corrugations, dots within the stimulus window translated horizontally in relation to the center (fixation) point with a sinusoidal velocity profile with a peak dot translation velocity (within the stimulus window) of 2.0 deg/sec. Therefore the stimulus had a maximum M/PRatio of ± 0.5. The direction of dot translation () in relation to direction of the stimulus window translation () determines the “depth phase” or the location of the “peaks” and “valleys” in the stimulus. That is, dots generating retinal image motion () in the opposite direction (image motion in the same direction) as dα are perceived near in depth, and depth phase can be varied independently of stimulus window translation (). The observer’s task was to report the depth phase of the stimulus (e.g., whether the top or bottom part of the stimulus appeared nearer than the fixation point).

The high-contrast pattern-masking stimulus was composed of 50 black and 50 white circles of varying sizes drawn upon a large (13 deg × 13 deg) black background. The circles were drawn, in alternating contrast, largest (7 deg) to smallest (2 deg) in random position upon the background. For each block of trials several different exemplars of the masking stimulus were created and used.

2.3. Participants

Five naïve observers performed a 2AFC discrimination of the depth phase of common random-dot motion parallax stimuli. All had normal or corrected-to-normal acuity. One observer completed six conditions and was excluded from participating in the last two conditions due to pain medication prescribed subsequent to orthopedic surgery. The medication severely affected this observer’s performance for one direction of eye movement, for an unusually long period of time, and her abnormal performance will be described in a subsequent report.

2.4. Procedure

The observer was seated with head supported in a chin rest. Each trial began with a fixation point at the center of the display. The observer initiated stimulus presentation with a key press. The fixation dot moved (laterally, left or right) immediately to the stimulus window starting position. Here the stimulus window would appear after a brief interval, randomly varying between 750 and 1500 msec. The stimulus window starting point was calculated so that the stimulus would appear moving across the center of the monitor, leftward moving stimuli started to the right of center, and right-ward moving stimuli started to the left of center. Stimuli having longer presentation durations started farther away from center.

The stimuli were presented in two interleaved staircases that varied the stimulus presentation duration for stimulus-window motion in leftward and rightward window movement. Both the leftward and rightward staircases started at 167 msec presentation duration and descended in 42 msec steps until its’ first reversal or the 42 msec total presentation duration. All subsequent staircase steps were in 8.3 msec intervals. The staircase decision rule tracked to 79% correct (Wetherill & Levitt, 1965) with each of the interleaved staircases ending with six reversals, and the observer’s performance determined from the average of the last four reversals.

Pursuit was monitored and enforced to a 1 deg window about the fixation point with the eye-tracking system. For each trial, stimulus presentation would only commence when the observer had established fixation on the stimulus fixation mark prior to initiating the trial. The 1 deg window was large enough, and stimulus translation speed low enough, to accommodate the initiation of pursuit without triggering a fixation loss. If a fixation loss was triggered anytime during a trial, then the stimulus was immediately extinguished, no observer response was recorded, and the trial was repeated. This enforcement procedure ensured that observers maintained fixation on, or very near, the fixation spot for each trial. Overall, only a small percentage of trials (< 2-3%) were repeated.

2.4.1. Experimental Conditions

Observers completed three of the interleaved staircases in each of the four experimental conditions, with and without the masking stimulus. In all four conditions observers reported perceived depth phase of the unambiguous motion parallax stimulus. The goal was to determine the minimum presentation interval required for an observer to accurately perform the depth phase (e.g., near-far) judgment. The pattern mask stimulus was used in “masking” conditions to determine whether the underlying visual processes required a longer presentation interval (SOA) than required in the non-masked condition. An initial pursuit interval (a pursuit prelude) was added before the presentation of the translating motion parallax stimulus to determine whether an ongoing pursuit signal would shorten the required presentation interval compared to conditions wherein observers programmed and initiated a pursuit eye movement concomitant with motion parallax stimulus presentation.

Condition 1: Depth

In this condition observers viewed only the translating motion parallax stimulus and reported on the perceived depth phase. Here observers maintained their gaze on the fixation spot at the center of the translating stimulus as it moved leftwards or rightwards.

Condition 2: Depth with Mask

In this condition the translating motion parallax stimulus (identical to condition 1) was followed (0 msec ISI) by the high-contrast pattern-masking stimulus. The pattern mask was centered upon the last fixation point and was about 4-times larger than the motion parallax stimulus. The pattern mask was presented for 83 msec.

Condition 3: Depth with Pursuit Prelude

In this condition the translating motion parallax stimulus was presented after a brief, variable (750 – 1500 msec) pursuit prelude during which observers viewed only a translating fixation target. After the variable pursuit interval, the translating motion parallax stimulus was presented along with the pursuit target, with the motion parallax stimulus presentation being identical to that of condition 1 and 2 above.

Condition 4: Depth with Mask and Pursuit Prelude

This condition included both the mask (condition 2) and pursuit prelude (condition 3). Trials began with the variable duration pursuit prelude (750-1500 msec), continued with translating motion parallax stimulus (identical to condition 1), and concluded with the pattern mask (0 msec ISI, 83 msec duration).

2.4.2. Control Conditions

A set of control conditions allows comparisons between the time required to perform a judgment of motion direction and a judgment of depth phase. Although retinal image motion is often unappreciated by the observer in conditions of motion parallax – objects generating retinal image motion are often perceived as stationary, not moving (e.g., see Ono & Ujike, 2005) – motion provides one of the crucial input signals to the motion parallax process. Therefore, it is important to determine presentation duration necessary for the visual system to determine the motion characteristics of the stimulus, in the absence of depth. Similar temporal intervals for the perception of motion and depth would suggest that the temporal processing limitations are due to the motion processing mechanism alone. One might expect identical temporal intervals for motion and depth if the perception of depth from motion parallax were based solely on retinal image motion. However, if the perception of depth requires a longer temporal interval, this result would suggest that motion processing is not the processing stage causing the temporal limitation. Instead, this would suggest that the perception of depth requires some additional time-consuming processing, such as the generation of an internal pursuit eye movement signal.

Condition 5: Stimulus Relative Motion

The goal of this condition was to determine the presentation duration required for observers to detect the directions of relative motion in the stimulus. In this case the stimulus window remained stationary in the center of the screen and observers performed judgments on the relative dot motion within the stimulus window. However, observers often reported perceiving some (ambiguous) depth in this stimulus, so the sine-wave modulation of dot motion was replaced by a square-wave modulation using the peak dot velocity of the sinusoidal stimulus(2.0 deg/sec). Observers reported the perceived direction of stimulus motion in a particular portion of the stimulus. In this condition observers maintained fixation on the stationary fixation point.

Condition 6: Stimulus Relative Motion with Mask

In this condition the relative motion stimulus (identical to condition 5) was followed (ISI 0 msec) with the pattern mask stimulus used in the experimental conditions 2 and 4 above. Again, the presentation duration of the pattern mask was 83 msec.

Condition 7: Stimulus Window Motion

In this condition the stimulus window, containing the relative motion stimulus from condition 5, translated leftward or rightward. Here the stimulus window started from the center position so that the starting point would not provide a cue to the direction of window movement. Here observers maintained fixation on the point at the center of the translating stimulus window.

Condition 8: Stimulus Window Motion with Mask

In this condition the standard pattern mask (0 msec ISI, 83 msec duration) was presented subsequent to the translating relative motion stimulus presentation (used in condition 7). Again, observers maintained fixation on the point at the center of the translating stimulus window.

3. Results

For each staircase a presentation duration threshold was calculated from the average of the last four reversals. For each condition an average threshold was calculated from the three different thresholds for each observer. Thresholds for leftward and rightward window motion were compared to determine whether observers showed a temporal difference for naso-temporal and temporo-nasal directions of eye movements. We found no evidence of a difference in normal observers. For example, the thresholds in the motion parallax (condition 1, no masking stimulus, no pursuit prelude) showed only a small difference between rightward (29.6 (0.49) msec) and leftward (27.2 (0.47) msec) window movement directions that was not statistically significant (p = 0.157). Therefore, thresholds for leftward and rightward window motion were combined to provide a single mean threshold for each condition. These means for each condition are shown in Figure 2.

Figure 2.

Figure 2

Average presentation durations (msec) required to accurately perform depth (top) and motion (bottom) discriminations. Error bars denote one standard error.

3.1 Depth from Motion Parallax (Experimental Conditions)

The perception of unambiguous depth from motion parallax can be achieved with very brief stimulus presentations of about 30 msec (Condition 1). Adding a 750 to 1500 msec pursuit “lead-in” did not affect the presentation duration required to correctly perceive depth from motion parallax (Condition 3). The presentation duration remained at about 30 msec. However, the presentation of a pattern mask had a large effect of the required presentation duration, regardless of whether the trials had a pursuit prelude (Condition 4) or not (Condition 2). The average presentation duration required to perform the depth phase discrimination was increased to 67 msec and 74 msec for these two conditions respectively.

3.2 Motion Perception (Control Conditions)

Observers can accurately report the direction of stimulus relative motion (Condition 5) and window motion (Condition 7) with very brief presentations of around 15 msec. Given the fixed frame rate of 120 Hz (8.3 msec per frame) this brief presentation duration corresponds to observers having little trouble performing these motion judgments with only 2 stimulus frames. The pattern mask did affect the presentation interval required to correctly perceive the direction of motion. It took about 35 msec to correctly judge the direction of relative dot motion within the stimulus (Condition 6) and about 25 msec to correctly judge the direction of stimulus window motion (Condition 8). These intervals required to determine motion direction are very similar to the intervals reported by Bischof and Di Lollo (1995) in their classic motion masking study.

4. Conclusion

A brief 30 msec presentation can be sufficient for an observer to unambiguously perceive depth from motion parallax. This indicates that the neural mechanisms serving the perception of depth from motion parallax are unexpectedly brisk compared to the “sluggish” estimates of close to a second of presentation time for depth from motion to “build up” (Eby, 1992; Hildreth et al., 1990; Andersen & Bradley, 1998, Treue et al., 1991). Such a brief integration time (≈ 30 msec) suggests that the ultimate precision of depth from motion parallax is unlikely to be limited predominately by the temporal properties of the neural processes. For example, consider that in 30 msec, a walking observer (5 km/hr) translates less than 5 cm. In this case, the true relative distance between two inline objects at 1 m and 1.25 m changes by less than a millimeter. However, in this example, observer translation over a period of 0.5 second would change the relative distance of the two objects by several centimeters. Given this change in the quantity being estimated over this long temporal interval, it is difficult to imagine how the visual system could ever generate an accurate estimate if the perceptual “build-up” time were close to one second.

The M/PLaw provides a precise mathematical framework within which to consider such implications of the current results. Consider the example given above, for two in-line objects at 1m and 1.25 m, using the M/PLaw (Formula 2) the estimate of depth changes only from the true relative distance of 0.25 m at t=0 to an estimate of 0.249 m at t=0.03 sec. If the 30 msec translation is centered on the point when the objects are in-line (t=-0.015 to t=+0.015) the depth estimate generates M/PL is between 0.2498 and 0.25 (see supplemental material for an interactive computation)1. In contrast, the M/PRatio (formula 1), which has the perceptual foreshortening of depth, generates relative depth estimates between 0.1999 m and 0.2 m during this 30 msec time interval for this particular viewing condition.

The reader can perform similar numerical examples for themselves (with a free browser plug-in) using interactive programs published at the Wolfram Demonstration Site, specifically Stroyan (2008.3; 2008.18). In the demonstration, one point can be made the fixate (1 m) and the other the distractor (1.25 m). Using the M/PLaw (formula 2), and M/PRatio (formula 1) this interactive demonstration shows how the relative depth between the fixate and the distractor changes as the observer translates. The reader can compare the estimate of d at two points of the translation by varying time, t, on the demonstration slider. These two demonstrations (Stroyan, 2008.3; 2008.18) assume a translation speed of 6.5 cm/sec, an approximate interocular translation distance in one second. This translation speed is slower than the examples above, but provides a useful way to understand how the relative depth between objects changes with observer translation. The crucial point of the demonstration is that a 30 msec temporal integration does not suggest a significant limitation to the potential accuracy of the motion/pursuit cue for depth perception.

While the perception of depth from motion parallax can be accurately recovered from a brief 30 msec presentation, the underlying neural processing could actually take a longer duration to complete. A better estimate of this is provided by the pattern mask conditions. In these masking conditions the required presentation duration for accurate depth perception is about 70-75 msec. This means that a high-contrast pattern mask can interrupt performance of the depth-phase discrimination task if presented sooner than 75 msec after stimulus presentation begins.

Is this because the pattern mask affects the perception of motion, which in turn affects the (Conditions 5-8) indicate the explanation. In masking conditions, the direction of relative stimulus that this is not motion () and the direction of the stimulus window movement (the proximal signal used by the visual system to generate ) can be accurately perceived with 15 msec presentations and 30 msec presentations. These results mean that the processing of visual motion has progressed beyond the effect of the pattern mask by about 30 msec. This is a much shorter presentation interval than that required to accurately perceived depth from motion parallax under the identical masking conditions. This means that the visual system has acquired both the translational and relative motion signals within the 30 msec presentation. However, this is insufficient for the system to generate the depth percept. This result is inconsistent with the idea that the perception of depth from motion parallax is based solely on motion processing. Additional processing, which can still be disrupted by the high-contrast pattern mask for the next 40-45 msec, must be occurring in the visual system.

We propose that this 70-75 msec interval is linked to the time course of generating the internal eye-movement signal (dα) necessary for the unambiguous perception of depth from motion parallax (Nawrot & Joyce, 2006; Nadler et al., 2009; Nawrot & Stroyan, 2009). If we consider information flow along the pursuit eye movement pathway (see Krauzlis, 2005; Kowler, 2011 for recent reviews), motion signals are processed quickly (30 msec), but additional stages such as attention (Bisley & Goldberg, 2003), target selection (Ferrera & Lisberger, 1997), and programming the eye-movement signal require additional time, leading to the large latencies to generate eye-movements. For instance, the ocular following response (OFR; Miles 1998) has latencies in the < 85 msec range, and pursuit initiation is typically 100 - 150 msec (Krauzlis, 2004, or even faster, Ferrera & Lisberger, 1995). Therefore, the 70-75 msec of processing time needed for this internal eye-movement signal to withstand disruption by the high-contrast pattern mask appears to be within the time period for the initiation of the OFR or pursuit.

Moreover, the high-contrast pattern mask could be disrupting the internal pursuit eye-movement signal in a variety of ways. One possible explanation is simple oculomotor capture (Theeuwes et al., 1998; Ludwig et al., 2008) where an abrupt stimulus onset (like the masking stimulus here) interferes with oculomotor programming. That is, the onset of the masking stimulus could “capture attention and evoke involuntary eye movements” (Lovejoy et al., 2009) that compete with the pursuit signal processing. Another possible explanation is suggested by Keller and Khan’s (1986) finding that a textured background (like the masking stimulus here) interferes with pursuit initiation, not just pursuit maintenance. Indeed, Kimmig et al. (1992) interpreted Keller and Khan’s results as, “Thus neither the eyes nor the retinal images of the background need to move for textured backgrounds to impair subsequent pursuit,…”. Therefore, it is possible that the masking stimulus in the present study impairs pursuit in a similar way.

However, it was unexpected that the pursuit prelude did not affect the 70-75 msec presentation SOA required to the high-contrast pattern mask (compare condition 4 to condition 2). Here the pursuit system was active for several hundred milliseconds prior to presentation of the translating motion parallax system, but the masking stimulus was still effective in disrupting the perception of depth for SOAs shorter than 70-75 msec. Knowing that the oculomotor system has a “velocity storage” mechanism, the hypothetical basis of anticipatory pursuit and other pursuit characteristics (Barnes & Donelan, 1999; Barnes & Collins, 2008), one might have expected that ongoing pursuit, initiated prior to presentation of the motion parallax stimulus, might have shortened the presentation interval required for accurate depth perception, prior to the mask. Nevertheless, the results show that the pursuit prelude had no effect. This result might be interpreted as meaning that the visual system does not use a “stored” pursuit velocity signal, and instead the depth perception process must receive an “active” an eye movement signal, generated in the same time frame as the relative retinal image motion. That is, it appears that the visual system recovers dα and dθ together, in the same time frame. This type of synchrony is a reasonable strategy for the visual system to prevent a perceptual error of having an “old” (and now incorrect in velocity or direction) stored eye-movement signal paired with a new retinal image motion.

Nevertheless, it is difficult to discount the alternative explanation is that within 70-75 msec the high-contrast pattern mask disrupts the overall “depth processing” mechanism, without specifically disrupting the generation of the internal pursuit signal. The specific parameters of this internal eye movement signal are difficult to measure directly because the observer’s eyes do not actually have to move to disambiguate depth from motion parallax. That is, the internal eye movement signal used for motion parallax is separable from a measureable eye movement. There are several demonstrations that the internal eye movement signal, generated to countermand a reflexive eye movement signal like TVOR (Nawrot, 2003; Nawrot & Joyce, 2006) and OKR (Nawrot & Stockert, 2005) and thereby maintain stable fixation, is sufficient to disambiguate motion parallax. Of course, in these cases the pursuit signal is in the direction opposite the TVOR or OKR, so the perceived depth is reversed (e.g. see Figure 11 in Nawrot & Joyce). All of these results show that an easily measureable external eye movement will not serve as a reliable proxy for the crucial internal eye movement signal.

One remaining question is: How does one reconcile the current study’s demonstration of very short presentations needed to correctly report unambiguous depth-sign from a motion parallax stimulus with the results of previous studies suggesting that the perception of structure from motion is slow, and builds up over time (Eby, 1992; Hildreth et al., 1990; Andersen & Bradley, 1998; Treue et al. 1991)? A recent analysis by Stroyan & Nawrot (2011) suggests an explanation based on the dynamic nature of the M/PRatio.

In the current experiment the corrugated motion parallax stimulus was constant over the horizontal axis, thereby making the fixate (F) and the relevant distractors (D, the maximum points of the peak and trough here) aligned along the naso-occipito axis. These conditions are well described by Figure 1 and the M/PLaw when t=0 (Nawrot & Stroyan, 2009).

However, stimuli in the previous studies, in which depth was found to “build-up”, had a complex 3-D shape that varied along the horizontal axis. The analysis of depth in such cases requires a more involved analysis (Stroyan & Nawrot, 2011) in which distractor points vary from the fixate in two dimensions. In the general case, the distractor (D) is displaced off the perpendicular line through the fixate (Figure 3). However, in these cases, the strict mathematical relation in formula (1) does not hold at t=0. This is because the M/PLaw has a time-varying characteristic that depends on relationship of D to the fixation point.

Figure 3.

Figure 3

Depiction of the geometry of motion parallax with the distractor displaced leftward from the line-of-sight position in the left panel. In this condition, the peak motion/pursuit ratio occurs when the hypothetical observer has moved to the right of the central position.

The motion/pursuit ratio, /dα[t], is a time-varying quantity that only gives a geometrically accurate cue to depth for certain points in space at certain times. Stroyan & Nawrot (2011) show that the maximum values of the motion/pursuit ratio can be geometrically aggregated to give an accurate representation of a 2-D shape (their Figure 21). Their Figure 20 shows the regions of space where this maximum is reached in three time intervals. The center panel shows a one second interval. Points in that part of space are accurately represented geometrically within that second, but points outside the region for the 30 msec interval on the left of Figure 20 take longer until the observer reaches a position where the cue is geometrically accurate. Their Figure 21 also shows the inaccuracy of the t=0 cue for a 2-D shape (a circle with a diameter with the t=0 cue on the left and the accurate aggregation on the right.) Therefore, it appears that the mathematically derived increase in the M/PRatio and the psychophysically determined build-up in the perception of depth from motion may share a very similar time course, depending on the location of objects in the scene.

Perhaps a few words of explanation are in order to explain why the maximum motion/pursuit cue could take longer for a complicated shape, although Stroyan & Nawrot’s (2011) results are geometrical and not empirical. The maximum of the motion/pursuit ratio occurs at the observer location where the relative depth is maximal. (This observation was not pointed out in Stroyan & Nawrot (2011), but they give a formula for the time when / is maximal and they occur at the same time.) Consider an observer shown in Figure 3 at t=0 on the left and a later time t > 0 on the right. If ρD[t] is the straight line (radial) distance from the eye to the distractor at time t and ρF[t] is the distance from the eye to the fixate, then (ρD - ρF )/ ρF is the relative depth of D beyond F. As the observer moves to the right from the t=0 position, the depth of D beyond F, (ρD - ρF ), gets bigger until D, F, and the observer’s eye are in line. However, the denominator term ρF in relative depth is also increasing, so tends to make the relative depth quantity decrease. The combined effect is that relative depth has a maximum at an observer position as shown in Figure 3 on the right, after the t=0 position, but before reaching the in-line position. This is the location where the motion/pursuit ratio for D is also maximal, giving the largest cue for D. As an observer moves, the “shape” of the 2D object builds up in time.

Therefore, one possible explanation for the longer integration times observed by (Hildreth, et al, 1990) and (Eby 1992) is that the more complicated set-ups in their experiments forced their observers to use the time-varying properties of the motion/pursuit ratio described in (Stroyan and Nawrot 2011). Their motions were different from ours, but if we consider a laterally translating observer (at speed 6.6 cm/sec) with fixate at 100 cm and distractor 20 cm deeper and 3 cm to the left of the fixate (a distance of 20.2 cm), the motion/pursuit ratio will not reach it’s maximum value until t=1.03 sec as shown on (Figure 4).2

Figure 4.

Figure 4

The time-varying (t) motion/pursuit law prediction of structural depth (d) of an object 20 cm deeper and 3 cm to the left of the fixate at 100 cm

While the visual system is sensitive to differences in the M/PRatio and perceives them as different depth magnitudes (Nawrot & Stroyan, 2009) it is unknown whether the visual system is sensitive to ongoing changes in M/PRatio, and is able to detect an increase (build-up) and peak (end of a build-up) in the M/PRatio. However, the similarity of the changes in the M/PRatio to the build-up in the perception of depth from motion parallax suggests that the relationship between the two will be an important avenue of psychophysical investigation.

Regardless of how the perception of depth from motion might build-up or change during observer translation, the results of the current study show that depth from motion parallax can be perceived with brief presentations of about 30 msec duration. This means that motion parallax can provide a quick source of relative depth information for an observer moving through a cluttered environment. Knowing the lower temporal integration limit for motion parallax provides a starting point to determine the ultimate accuracy of depth from motion parallax for various viewing conditions.

Supplementary Material

01

Highlights.

  • -perception of depth from motion parallax occurs with brief stimulus presentations (30 msec)

  • -a high-contrast pattern mask can interrupt this with an SOA less than 70-75 msec

  • -the masking interval might be tied to the latency of eye movement proessing

  • -mechanisms serving the perception of depth from motion parallax are not necessarily sluggish

Acknowledgements

This work was supported by a Centers of Biomedical Research Excellence (COBRE) grant: NIH P20 RR020151. The National Center for Research Resources (NCRR) is a component of the National Institutes of Health (NIH). The contents of this report are solely the responsibility of the authors and do not necessarily reflect the official views of the NIH or NCRR.

Footnotes

1

The supplemental materials contain an interactive program that can be used with the Wolfram CDF Player, available free at: http://www.wolfram.com/products/player/ The “.cdf” file suffix represents a document type that stands for “computational document format.” These documents need a “reader” similar to the Adobe Reader for “.pdf” documents. Once you download and install the Wolfram CDF Player, you will be able to perform interactive computations with the document, similar to the walking example outlined above. The supplemental program uses the specific parameters of our experiment such as the f=57 cm fixate distance and our 4 cm/sec translation speed. At t=100 msec the M/PLaw computation of a distractor with d=40 only differs in the fourth digit from the maximal value at t=0 (0.7017 vs 0.7018).

2

The .cdf attachment in the supplemental materials (described above for computing the M/P Ratio) has a second part. For rough comparison to Eby, we used a parabolic shape corresponding to the horizontal cross section of his stimulus. Our translation is linear, rather than circular, but about the same size – and the size can be varied interactively. The interactive program shows the times and positions where the maximal predictions are reached for a number of points on the arcs and tracks 6 of those points.

Publisher's Disclaimer: This is a PDF file of an unedited manuscript that has been accepted for publication. As a service to our customers we are providing this early version of the manuscript. The manuscript will undergo copyediting, typesetting, and review of the resulting proof before it is published in its final citable form. Please note that during the production process errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain.

References

  1. Andersen RA, Bradley DC. Perception of three-dimensional structure from motion. Trends in Cognitive Science. 1998:222–228. doi: 10.1016/s1364-6613(98)01181-4. [DOI] [PubMed] [Google Scholar]
  2. Barnes GR, Donelan AS. The remembered pursuit task: Evidence for segregation of timing and velocity storage in predictive oculomotor control. Experimental Brain Research. 1999;129:57–67. doi: 10.1007/s002210050936. [DOI] [PubMed] [Google Scholar]
  3. Barnes GR, Collins CJS. Evidence for a link between the extra-retinal component of random-onset pursuit and the anticipatory pursuit of predictable object motion. Journal of Neurophysiology. 2008;100:1135–1146. doi: 10.1152/jn.00060.2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Bischof WF, Di Lollo V. On motion and metacontrast with simultaneous onset of the stimuli. Journal of the Optical Society of America A. 1995;12:1623–1636. doi: 10.1364/josaa.12.001623. [DOI] [PubMed] [Google Scholar]
  5. Bisley JW, Goldberg ME. Neuronal activity in the lateral interparietal area and spatial attention. Science. 2003;299:81–86. doi: 10.1126/science.1077395. [DOI] [PubMed] [Google Scholar]
  6. Breitmeyer BG. Visual masking: An integrative approach. Oxford University Press; New York: 1984. [Google Scholar]
  7. Domini F, Caudek C. Perceiving surface slant from deformation of optic flow. Journal of Experimental Psychology: Human Perception and Performance. 1999;25:426–444. doi: 10.1037//0096-1523.25.2.426. [DOI] [PubMed] [Google Scholar]
  8. Eby DW. The spatial and temporal characteristics of perceiving 3-D structure from motion. Perception & Psychophysics. 1992;51:163–78. doi: 10.3758/bf03212240. [DOI] [PubMed] [Google Scholar]
  9. Fermüller C, Aloimonos Y. On the geometry of visual correspondence. International Journal of Computer Vision. 1997;21:223–247. [Google Scholar]
  10. Ferrera VP, Lisberger SG. Attention and target selection for smooth pursuit eye movements. Journal of Neuroscience. 1995;15:7472–7484. doi: 10.1523/JNEUROSCI.15-11-07472.1995. [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Ferrera VP, Lisberger SG. Neuronal responses in visual areas MT and MST during smooth pursuit target selection. Journal of Neurophysiology. 1997;78:1433–1446. doi: 10.1152/jn.1997.78.3.1433. [DOI] [PubMed] [Google Scholar]
  12. Freeman TCA, Fowler TA. Unequal retinal and extra-retinal motion signals produce different perceived slants of moving surfaces. Vision Research. 2000;40:1857–1868. doi: 10.1016/s0042-6989(00)00045-6. [DOI] [PubMed] [Google Scholar]
  13. Gordon DA. Static and dynamic fields in human space perception. Journal of the Optical Society of America. 1965;55:1296–1303. doi: 10.1364/josa.55.001296. [DOI] [PubMed] [Google Scholar]
  14. Hanes DA, Keller J, McCollum G. Motion parallax contribution to perception of self-motion and depth. Biological Cybernetics. 2008;98:273–293. doi: 10.1007/s00422-008-0224-2. [DOI] [PubMed] [Google Scholar]
  15. Hildreth EC, Grzywacz NM, Adelson EH, Inada VK. The perceptual buildup of three-dimensional structure from motion. Perception & Psychophysics. 1990;48:19–36. doi: 10.3758/bf03205008. [DOI] [PubMed] [Google Scholar]
  16. Keller EL, Khan NS. Smooth pursuit initiation in the presence of a textured background in monkey. Vision Research. 1986;26:943–955. doi: 10.1016/0042-6989(86)90152-5. [DOI] [PubMed] [Google Scholar]
  17. Kim HR, Angelaki DE, DeAngelis GC. 2011 Neuroscience Meeting Planner. Society for Neuroscience; Washington, DC: 2011. Neural correlates of depth perception from motion parallax in macaque area MT. Program No. 577.08. 2011. Online. [Google Scholar]
  18. Kimmig HG, Miles FA, Schwarz U. Effects of stationary textured backgrounds on the initiation of pursuit eye movements in monkeys. Journal of Neurophysiology. 1992;68:2147–2164. doi: 10.1152/jn.1992.68.6.2147. [DOI] [PubMed] [Google Scholar]
  19. Koenderink JJ, van Doorn AJ. Invariant properties of the motion parallax field due to the movement of rigid bodies relative to the observer. Optic Acta. 1976;22:773–779. [Google Scholar]
  20. Koenderink JJ, van Doorn AJ. Facts on optic flow. Biolological Cybernetics. 1987;56:247–254. doi: 10.1007/BF00365219. [DOI] [PubMed] [Google Scholar]
  21. Kowler E. Eye movements: The past 25 years. Vision Research. 2011;51:1457–1483. doi: 10.1016/j.visres.2010.12.014. [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Krauzlis RJ. Recasting the smooth pursuit eye movement system. Journal of Neurophysiology. 2004;91:591–603. doi: 10.1152/jn.00801.2003. [DOI] [PubMed] [Google Scholar]
  23. Krauzlis RJ. The control of voluntary eye movements: New perspectives. Neuroscientist. 2005;11:124–137. doi: 10.1177/1073858404271196. [DOI] [PubMed] [Google Scholar]
  24. Lappe M, Bremmer F, van den Berg AV. Perception of self-motion from visual flow. s. 1999. pp. 329–336. [DOI] [PubMed]
  25. Longuet-Higgins HC, Prazdny K. The interpretation of a moving retinal image. Proceedings of the Royal Society London. 1980;208:385–397. doi: 10.1098/rspb.1980.0057. [DOI] [PubMed] [Google Scholar]
  26. Lovejoy LP, Fowler GA, Krauzlis RJ. Spatial allocation of attention during smooth pursuit eye movements. Vision Research. 2009;49:1275–1285. doi: 10.1016/j.visres.2009.01.011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  27. Ludwig CJH, Ranson A, Gilchrist ID. Oculomotor capture by transient events: A comparison of abrupt onsets, offsets, motion and flicker. Journal of Vision. 2008;8:1–16. doi: 10.1167/8.14.11. [DOI] [PubMed] [Google Scholar]
  28. Miles FA. The neural processing of 3-D visual information: evidence from eye movements. European Journal of Neuroscience. 1998;10:811–822. doi: 10.1046/j.1460-9568.1998.00112.x. [DOI] [PubMed] [Google Scholar]
  29. Moeeny A, Cumming B. 2011 Neuroscience Meeting Planner. Society for Neuroscience; Washington, DC: 2011. Effect of pursuit eye movement on responses of MT neurons systematically depends on disparity and motion direction preference. Program No. 578.12. 2011. Online. [Google Scholar]
  30. Nadler JW, Angelaki DE, DeAngelis GC. A neural representation of depth from motion parallax in macaque visual cortex. Nature. 2008;452:642–645. doi: 10.1038/nature06814. [DOI] [PMC free article] [PubMed] [Google Scholar]
  31. Nadler JW, Nawrot M, Angelaki DE, DeAngelis GC. MT neurons combine visual motion with a smooth eye movement signal to code depth sign from motion parallax. Neuron. 2009;63:523–532. doi: 10.1016/j.neuron.2009.07.029. [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Naji JJ, Freeman TCA. Perceiving depth order during pursuit eye movements. Vision Research. 2004;44:3025–3034. doi: 10.1016/j.visres.2004.07.007. [DOI] [PubMed] [Google Scholar]
  33. Nakayama K, Loomis JM. Optical velocity patterns, velocity-sensitive neurons, and space perception: A hypothesis. Perception. 1974;3:63–80. doi: 10.1068/p030063. [DOI] [PubMed] [Google Scholar]
  34. Nawrot M. Eye movements provide the extra-retinal signal required for the perception of depth from motion parallax. Vision Research. 2003;43:1553–1562. doi: 10.1016/s0042-6989(03)00144-5. [DOI] [PubMed] [Google Scholar]
  35. Nawrot M, Joyce L. The pursuit theory of motion parallax. Vision Research. 2006;46:4709–4725. doi: 10.1016/j.visres.2006.07.006. [DOI] [PubMed] [Google Scholar]
  36. Nawrot M, Stockert C. Motion parallax in movies: Background motion, eye movement signals, and depth. Journal of Vision (abstract) 2005;5(8):644. [Google Scholar]
  37. Nawrot M, Stroyan K. The motion/pursuit law for visual depth perception from motion parallax. Vision Research. 2009;49:1969–1978. doi: 10.1016/j.visres.2009.05.008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  38. Ono H, Ujike H. Motion parallax driven by head movements: Conditions for visual stability, perceived depth, and perceived concomitant motion. Perception. 2005;34:477–490. doi: 10.1068/p5221. [DOI] [PubMed] [Google Scholar]
  39. Perrone JA, Stone LS. A model of self-motion estimation within primate extrastriate visual cortex. Vision Research. 1994;43:2917–2938. doi: 10.1016/0042-6989(94)90060-4. [DOI] [PubMed] [Google Scholar]
  40. Stroyan KD. Interactive computation of Geometric Inputs to Visual Depth Perception. 2008 index at http://www.math.uiowa.edu/~stroyan/VisionWebsite/PlayOutline.html.
  41. Motion Pursuit Law In 1D Visual Depth Perception 1. 2008.1 http://demonstrations.wolfram.com/MotionPursuitLawIn1DVisualDepthPerception1/
  42. Motion Pursuit Fixate And Distraction: Visual Depth Perception 2. 2008.2 http://demonstrations.wolfram.com/MotionPursuitFixateAndDistractorVisualDepthPerception2/
  43. Motion Pursuit Law In 2D: Visual Depth Perception 3. 2008.3 http://demonstrations.wolfram.com/MotionPursuitLawIn2DVisualDepthPerception3/
  44. Motion Pursuit Ratio And Depth In 1D: Visual Depth Perception 17. 2008.17 http://demonstrations.wolfram.com/MotionPursuitRatioAndDepthIn1DVisualDepthPerception17/
  45. Motion Pursuit Ratio And Depth In 2D: Visual Depth Perception 18. 2008.18 http://demonstrations.wolfram.com/MotionPursuitRatioAndDepthIn2DVisualDepthPerception18/
  46. Stroyan K, Nawrot M. Visual depth from motion parallax and eye pursuit. Journal of Mathematical Biology. 2011 doi: 10.1007/s00285-011-0445-1. in press. DOI 10.1007/s00285-011-0445-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  47. Theeuwes J, Kramer AF, Hahn S, Irwin DE. Our eyes do not always go where we want them to go: Capture of the eyes by new objects. Psychological Science. 1998;9:379–385. [Google Scholar]
  48. Treue S, Husain H, Andersen RA. Human perception of structure from motion. Vision Research. 1991;31:59–75. doi: 10.1016/0042-6989(91)90074-f. [DOI] [PubMed] [Google Scholar]
  49. von Helmholtz H. Treatise on Physiological Optics. Dover; New York: 1910/1925/1962. 1962. (English translation by Southall JPC, three volumes bound as two, from the 3rd German edition of Handbuch der Physiologischen Optik. von Kries Note 4 appeared in the 1925 edition.) [Google Scholar]
  50. Wallach H, O’Connell DN. The kinetic depth effect. The Journal of Experimental Psychology. 1953;45:205–217. doi: 10.1037/h0056880. [DOI] [PubMed] [Google Scholar]
  51. Wetherill GB, Levitt H. Sequential estimation of points on a psychometric function. The British Journal of Mathematical and Statistical Psychology. 1965;18:1–10. doi: 10.1111/j.2044-8317.1965.tb00689.x. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

01

RESOURCES