Skip to main content
UKPMC Funders Author Manuscripts logoLink to UKPMC Funders Author Manuscripts
. Author manuscript; available in PMC: 2010 Mar 12.
Published in final edited form as: J Vis. 2009 Dec 9;9(13):11.1–1137. doi: 10.1167/9.13.11

Latitude and longitude vertical disparity

Jenny C A Read 1, Graeme P Phillipson 2, Andrew Glennerster 3
PMCID: PMC2837276  EMSID: UKMS28470  PMID: 20055544

Abstract

The literature on vertical disparity is complicated by the fact that several different definitions of the term “vertical disparity” are in common use, often without a clear statement about which is intended or a widespread appreciation of the properties of the different definitions. Here, we examine two definitions of retinal vertical disparity: elevation-latitude and elevation-longitude disparity. Near the fixation point, these definitions become equivalent, but in general, they have quite different dependences on object distance and binocular eye posture, which have not previously been spelt out. We present analytical approximations for each type of vertical disparity, valid for more general conditions than previous derivations in the literature: we do not restrict ourselves to objects near the fixation point or near the plane of regard, and we allow for non-zero torsion, cyclovergence and vertical misalignments of the eyes. We use these expressions to derive estimates of the latitude and longitude vertical disparity expected at each point in the visual field, averaged over all natural viewing. Finally, we present analytical expressions showing how binocular eye position – gaze direction, convergence, torsion, cyclovergence, and vertical misalignment – can be derived from the vertical disparity field and its derivatives at the fovea.

Keywords: binocular vision, stereopsis, depth perception, induced effect, vertical disparity, computational vision

Introduction

Because the two eyes are set apart in the head, the images of an object fall at different positions in the two eyes. The resulting binocular disparity is, in general, a two-dimensional, vector quantity. Since Helmholtz (1925), psychophysicists have divided this vector into two components: horizontal and vertical disparity. There is now an extensive literature discussing how vertical disparities influence perception and how this may be achieved within the brain, for example (Backus, Banks, van Ee, & Crowell, 1999; Banks, Backus, & Banks, 2002; Berends, van Ee, & Erkelens, 2002; Brenner, Smeets, & Landy, 2001; Cumming, 2002; Duke & Howard, 2005; Durand, Zhu, Celebrini, & Trotter, 2002; Friedman, Kaye, & Richards, 1978; Frisby, et al., 1999; Garding, Porrill, Mayhew, & Frisby, 1995; Gillam & Lawergren, 1983; Kaneko & Howard, 1997b; Longuet-Higgins, 1982; Matthews, Meng, Xu, & Qian, 2003; Ogle, 1952; Read & Cumming, 2006; Rogers & Bradshaw, 1993, 1995; Westheimer, 1978; Williams, 1970). Yet the literature is complicated by the fact that the term “vertical disparity” is used in several different ways by different authors. The first and most fundamental distinction is whether disparity is defined in a head-centric or retino-centric coordinate system. The second issue concerns how disparity, as a two-dimensional vector quantity, is divided up into “vertical” and “horizontal” components.

In a head-centric system, vertical and horizontal disparities are defined in the optic array, that is the set of light rays crossing the nodal points of each eye. One chooses an angular coordinate system to describe the line of sight from each eye to a point in space; vertical disparity is then defined as the difference in the elevation coordinates. If Helmholtz coordinates are used (Figure 1A, (Howard & Rogers, 2002)), then for any point in space, the elevation is the same from both eyes. Thus, head-centric Helmholtz elevation disparity is always zero for real objects (Erkelens & van Ee, 1998). This definition is common in the physiology literature, where “vertical disparity” refers to a vertical off-set between left and right images on a frontoparallel screen (Cumming, 2002; Durand, Celebrini, & Trotter, 2006; Durand, et al., 2002; Gonzalez, Justo, Bermudez, & Perez, 2003; Stevenson & Schor, 1997). In this usage, a “vertical” disparity is always a “non-epipolar” disparity; that is, a two-dimensional disparity which cannot be produced by any real object, given the current eye position. A different definition uses Fick coordinates to describe the angle of the line of sight (Figure 1B). With this definition, vertical disparities occur in natural viewing (e.g. (Backus & Banks, 1999; Backus, et al., 1999; Bishop, 1989; Hibbard, 2007; Rogers & Bradshaw, 1993)), and so non-zero vertical disparities are not necessarily non-epipolar.

Figure 1.

Figure 1

Two coordinates systems for describing head-centric or optic-array disparity. Red lines are drawn from the two nodal points L, R to an object P. A: Helmholtz coordinates. Here, we first rotate up through the elevation angle λ to get us into the plane LRP, and the azimuthal coordinate ζ rotates the lines within this plane until they point to P. The elevation is thus the same for both eyes; no physical object can have a vertical disparity in optic-array Helmholtz coordinates. B: Fick coordinates. Here, the azimuthal rotation ζ is applied within the horizontal plane, and the elevation λ then lifts each red line up to point at P. Thus, elevation is in general different for the two lines, meaning that the object P has a vertical disparity in optic-array Fick coordinates.

Head-centric disparity is independent of eye position, and thus mathematically tractable. However, all visual information arrives on the retinas, and it seems clear that the brain's initial encoding of disparity is retinotopic. Accordingly, the appropriate language for describing the neuronal encoding of disparity must be retinotopic (Garding, et al., 1995). The retinal disparity of an object is the two-dimensional vector linking its two images in the two retinas. This depends both on the scene viewed, and also on the binocular posture of the eyes viewing it. Thus, two-dimensional retinal disparity can be used to extract eye posture as well as scene structure (Garding, et al., 1995; Longuet-Higgins, 1981). It is more complicated to handle than head-centric disparity, but it contains more information.

For retinal disparity, as well, two definitions of vertical disparity are in common use within the literature, stemming from the difficulty of defining a “vertical” direction on a spherical eyeball. One possibility is to define “vertical” as the projection of vertical lines in space onto the retina, so that a line of constant “vertical” position on the retina is the projection of a horizontal line in space. This produces lines of elevation-longitude on the spherical eyeball (Figure 2BD; Figure 3AC). We shall refer to the corresponding vertical coordinate as elevation longitude, η; it is called inclination by Bishop (1962). Equivalently, one can project the hemispherical retina onto a plane, and take the vertical Cartesian coordinate on the plane, y (Figure 2A). This is usual in the computer vision literature. Since there is a simple one-to-one mapping between these two coordinates, y=tanη, we shall not need to distinguish between them in this paper. The natural definition of vertical disparity within this coordinate system is then the difference between the elevation-longitude of the images in the two eyes. Papers that have defined vertical disparity to be differences in either η or y include (Garding, et al., 1995; Hartley & Zisserman, 2000; Longuet-Higgins, 1982; Mayhew, 1982; Mayhew & Longuet-Higgins, 1982; Read & Cumming, 2006).

Figure 2.

Figure 2

Different retinal coordinate systems. A: Cartesian planar. Here, the x and y refer to position on the virtual plane behind the retina; the “shadow” shows where points on the virtual plane correspond to on the retina, i.e. where a line drawn from the virtual plane to the center of the eyeball intersects the eyeball. B: Azimuth-longitude/elevation-longitude. C: Azimuth-longitude/elevation-latitude. D: Azimuth-latitude/elevation-longitude. E: Azimuth-latitude/elevation-latitude. For the angular coordinate systems (B-E), lines of latitude/longitude are drawn at 15° intervals between ±90°. For the Cartesian system (A), the lines of constant x and y are at intervals of 0.27=tan15°. Lines of constant x are also lines of constant α, but lines that are equally spaced in x are not equally spaced in α.

Figure 3.

Figure 3

Two definitions of vertical retinal disparity. AB) A point in space, P, projects to different positions IL and IR on the two retinae. CD) The two retinae are shown superimposed, with the two half-images of P shown in red and blue for the left and right retinae respectively. In (AC), the retinal coordinate system is azimuth-longitude/elevation-longitude. In (BD), it is azimuth-longitude/elevation-latitude. The point P and its images IL and IR are identical between (AC) and (BD); the only difference between left and right halves of the figure is the coordinate system drawn on the retinae. The eyes are converged 30° fixating a point on the midline: X=0, Y=0, Z=11. The plane of gaze, the XZ plane, is shown in gray. Lines of latitude and longitude are drawn at 15° intervals. The point P is at X=−6, Y=7, Z=10. In elevation-longitude coordinates, the images of P fall at ηL=−30°, ηR=−38°, so the vertical disparity ηΔ is −8°. In elevation-latitude, κL=−27°, κR=−34°, and the vertical disparity κΔ=−6°. This figure was generated by Fig_VDispDefinition.m in the Supplementary Material.

An alternative approach is to define the vertical coordinate as being lines of latitude on the sphere (Figure 2CE; Figure 3BD). We shall refer to the corresponding vertical coordinate as elevation latitude, κ. (See Table 3 in Appendix A for a complete list of symbols used in this paper). Studies defining vertical disparity as the difference in elevation-latitude κ include (Barlow, Blakemore, & Pettigrew, 1967; Bishop, et al., 1962; Howard & Rogers, 2002).

Table 3.

Difinition of symbols.

I Interocular distance
i half interocular distance, i=I/2
k,l integer counters taking on values 1, 2, 3.
ML,MR rotation matrix for left, right eye respectively
Mc cyclopean rotation matrix, Mc=(MR+ML)/2
Mδ half-difference rotation matrix, Mδ=(MR−ML)/2
m vectors mj are the three columns of the corresponding rotation matrix M:
E.g., mc1 = [Mc11 Mc21 Mc31] ; mδ2 = [Mδ12 Mδ22 Mδ32] (Equation 33)
HL,R,c gaze azimuth in Helmholtz system for left, right, cyclopean eye
VL,R,c gaze elevation in Helmholtz system for left, right, cyclopean eye
TL,R,c gaze torsion in Helmholtz system for left, right, cyclopean eye
HΔ horizontal vergence
VΔ vertical vergence (non-zero values indicate a failure of fixation)
TΔ cyclovergence
X,Y,Z position in space in Cartesian coordinates fixed with respect to the head (Figure 12)
unit vector parallel to the X-axis.
P vector representing position in space in head-centered coordinates: P=(X,Y,Z)
U,W,S position in space in Cartesian coordinates fixed with respect to the cyclopean gaze. The S axis is the
optic axis of the cyclopean eye (see Figure 5).
R distance of an object from the origin. R2=X2+Y2+Z2= U2+W2+S2 (see Figure 5).
R0 distance of the fixation point from the origin (or distance to the point where the gaze rays most nearly
intersect, if the eyes are misaligned so that no exact intersection occurs).
δ fractional difference between the fixation distance, R0, and the distance to the object under consideration,
R. That is, δ = (R−R0)/R0
x horizontal position on the retina in Cartesian coordinate system (Figure 2A)
y vertical position on the retina in Cartesian coordinate system (Figure 2A)
α azimuth-longitude coordinate for horizontal position on the retina (Figure 2BC)
η elevation-longitude coordinate for vertical position on the retina (Figure 2BD)
β azimuth-latitude or declination, coordinate for horizontal position on the retina (Figure 2DE)
κ elevation-latitude or inclination, coordinate for vertical position on the retina (Figure 2CE)
ξ Retinal eccentricity (Equation 10)

Both definitions of vertical disparity are perfectly valid and in common use. The trouble is that statements that are true of one are not true of the other. For example, elevation-longitude vertical disparity is always zero when the eyes are in primary position, but elevation-latitude vertical disparity is not. Horizontal rotations of the eyes away from primary position change the elevation-longitude to which an object projects, but leave its elevation-latitude unaltered. As a consequence, elevation-latitude vertical disparity is independent of convergence, whereas elevation-longitude vertical disparity increases as the eyes converge. Elevation-latitude vertical disparity is always zero for objects on the mid-sagittal plane, but elevation-longitude vertical disparity is not. Yet despite these crucial differences, papers on vertical disparity often do not spell out which definition they are employing. From personal experience, we believe that the differences between the definitions are not widely appreciated, perhaps because the two definitions become equivalent at the fovea. A key aim of this paper is to lay out the similarities and differences between both definitions in a single convenient reference.

A second aim is to obtain analytical expressions for both types of disparity that are as general as possible. Most mathematical treatments in the psychology literature make simplifying assumptions, e.g. that the eyes have no torsion, that the object is in the plane of gaze, that the eyes are correctly fixating on a single point in space. Conversely the computer vision literature allows for completely general camera positions, but does not provide explicit expressions for vertical disparity. Here, we derive unusually general, explicit expressions for both types of vertical disparity. We still use the small baseline approximation required by previous treatments, and also small vergence angles. However, we are able to produce approximate expressions that allow for small amounts of cyclotorsion, cyclovergence and vertical misalignments between the eyes.

These general expressions allow us to derive simple expressions for the expected pattern of vertical disparity across the visual field, averaged across all scenes and eye positions. A few previous studies have attempted to estimate the two-dimensional distribution of disparities encountered in normal viewing (Hibbard, 2007; Y. Liu, Bovik, & Cormack, 2008; Read & Cumming, 2004), but these have all averaged results across the entire visual field. This study is the first to examine the expected vertical disparity as a function of position in the visual field, and we hope it will be useful to physiologists studying the neuronal encoding of vertical disparity.

In the vicinity of the fovea, the distinction between latitude and longitude becomes immaterial. The two definitions of vertical disparity are therefore equivalent. We show how eye position can be read off very straightforwardly from this unified vertical disparity. We derive simple analytic expressions giving estimates of each eye position parameter in terms of the vertical disparity at the fovea, and its rate of change there. These are almost all implicit in the existing literature (Backus & Banks, 1999; Backus, et al., 1999; Banks, et al., 2002; Banks, Hooge, & Backus, 2001; Kaneko & Howard, 1996, 1997b; Koenderink & van Doorn, 1976 1478; Rogers & Bradshaw, 1993, 1995; Rogers & Cagenello, 1989), but are brought together here within a single clear set of naming conventions and definitions so that the similarities and differences between definitions can be readily appreciated.

Results

1. General expressions for elevation-longitude and elevation-latitude vertical disparity

Figure 3 shows the two definitions of retinal vertical disparity that we consider in this paper. A point P in space is imaged to the points IL and IR in the left and right retinae respectively (Figure 3AB). Figure 3C and D show the left and right retinae aligned and superimposed, so that the positions of the images IL and IR can be more easily compared. The left panels (AC) and right-hand panels (BD) of Figure 3 are identical apart from the vertical coordinate system drawn on the retina: Figure 3AC shows elevation-longitude η, and Figure 3BD shows elevation-latitude κ. The vertical disparity of the point P is the difference between the vertical coordinates of its two half-images. For this example, the elevation-longitude vertical disparity of P is ηΔ = −8°, while the elevation-latitude disparity is κΔ=−6°.

In the Appendices, we derive approximate expressions for both types of retinal vertical disparity. These are given in Table 7. In general, vertical disparity depends on the position of object P (both its visual direction and its distance from the observer), and on the binocular posture of the eyes. Each eye has three degrees of freedom, which we express in Helmholtz coordinates as the gaze azimuth H, elevation V and torsion T (Figure 4; Appendix A). Thus, in total the two eyes have potentially 6 degrees of freedom. It is convenient to represent these by the mean and the difference between the left and right eyes. Thus, we shall parametrize eye position by the three coordinates of an imaginary cyclopean eye (Figure 5), Hc, Vc and Tc, and the three vergence angles, HΔ, VΔ and TΔ, where Hc=(HR+HL)/2, and HΔ=HR−HL, etc (Table 2, Table 3). When we refer below to convergence, we mean the horizontal vergence angle HΔ. We shall refer to VΔ as vertical vergence error or vertical vergence misalignment. We call VΔ a misalignment because, in order for the two eyes' optic axes to intersect at a single fixation point, VΔ must be zero, and this is empirically observed to be usually the case. We shall call TΔ the cyclovergence. Non-zero values of TΔ mean that the eyes have rotated in opposite directions about their optic axes. This occurs when the eyes look up or down: when the eyes look up, they also rotate about their lines of sight, such that the top of each eyeball moves nearer the nose. Cyclovergence in the opposite direction occurs when the eyes look down. We shall refer to Tc as cyclopean torsion or cyclotorsion. Non-zero values of Tc mean that the two eyes are both rotated in the same direction. This happens, for example, when the head tilts to the left; both eyes counter-rotate slightly in their sockets so as to reduce their movement in space, i.e. anti-clockwise as viewed by someone facing the observer (Carpenter, 1988).

Table 7.

Expressions for vertical disparity in different coordinate systems. These are correct to first-order in interocular distance I/S (I/R) and in the vergence angles HΔ, VΔ, and TΔ. They hold all over the retina and for any cyclopean gaze Hc, elevation Vc or overall cyclotorsion Tc.

Vertical disparity Most general expressions
In planar Cartesian
retinal coordinates as
a function of spatial
position in gaze-
centered coordinates
yΔIS(WSHcsin+TcsinHccos)+[UWS2Tccos(W2S2+1)Tcsin]HΔ[USHcsin(W2S2+1)HccosTccosUWS2HccosTcsin]VΔ+USTΔ
In planar Cartesian
retinal coordinates as
a function of retinal
location in planar
Cartesian coordinates
yΔ(ycHcsinTcsinHccos)IS+[xcycTccos(yc2+1)Tcsin]HΔ[xcHcsin+(yc2+1)HccosTccos+xcycHccosTcsin]VΔxcTΔ
In elevation-
longitude, as a function
of spatial location
in gaze-centered
coordinates
ηΔ1W2+S2{[WHcsin+STcsinHccos]I+[(W2+S2)Tcsin+UWTccos]HΔ[(W2+S2)HccosTccos+UWHccosTcsinUSHcsin]VΔ+USTΔ}
In elevation-
longitude, as a function
of retinal location
in azimuth-
longitude/elevation-
longitude coordinates
ηΔcos2ηc[TcsinHccosHcsintanηc]IS+[tanαcsinηccosηcTccosTcsin]HΔ[HccosTccos+tanαcsinηccosηcHccosTcsin+tanαccos2ηcHcsin]VΔtanαccos2ηcTΔ
In elevation -
longitude, as a function
of retinal location
in azimuth-
latitude/elevation-
longitude coordinates
ηΔcos2ηc[TcsinHccosHcsintanηc]IS+[tanβcsinηcTccosTcsin]HΔ[HccosTccos+tanβcsinηcHccosTcsin+tanβccosηcHcsin]VΔtanβccosηcTΔηΔIRcosηccosβc[TcsinHccosHcsintanηc]+[tanβcsinηcTccosTcsin]HΔ[HccosTccos+tanβcsinηcHccosTcsin+tanβccosηcHcsin]VΔtanβccosηcTΔ
In elevation-latitude,
as a function of spatial
location in gaze-
centered coordinates
κΔ=I(HccosTccosUW+HcsinWS+TcsinHccos(U2+S2))(U2+W2+S2)U2+S2(STcsinHΔ+SHccosTccosVΔUHcsinVΔUTΔ)U2+S2
In elevation-latitude,
as a function of retinal
location in azimuth-
longitude
/elevation- latitude
coordinates
κΔ=IScosαccosκc(TccosHccossinαcsinκcHcsincosαcsinκc+TcsinHccoscosκc)TcsinHΔcosαc(HccosTccoscosαc+sinαcHcsin)VΔsinαcTΔκΔ=IR(TccosHccossinαcsinκcHcsincosαcsinκc+TcsinHccoscosκc)TcsinHΔcosαc(HccosTccoscosαc+sinαcHcsin)VΔsinαcTΔ

Figure 4.

Figure 4

Helmholtz coordinates for eye position, (a) shown as a gimbal, after Howard (2002) Fig. 9.10, (b) shown for the cyclopean eye. The sagittal YZ plane is shown in blue, the horizontal XZ plane in pink, and the gaze plane in yellow. There are two ways of interpreting Helmholtz coordinates: (1) Starting from primary position, the eye first rotates through an angle T about an axis through the nodal point parallel to Z, then through H about an axis parallel to Y, and finally through V about an axis parallel to X. Equivalently (2), starting from primary position, the eye first rotates downwards through V, bringing the optic axis into the desired gaze plane (shown in yellow) then rotates through H about an axis orthogonal to the gaze plane, and finally through T about the optic axis.

Figure 5.

Figure 5

Different ways of measuring the distance to the object P. The two physical eyes are shown in gold; the cyclopean eye is in between them, in blue. F is the fixation point; the brown lines mark the optic axes, and the blue line the direction of cyclopean gaze. The point P is marked with a red dot. It is at a distance R from the origin. Its perpendicular projection on the cyclopean gaze axis is also drawn in red (with a corner indicating the right-angle); the distance of this projection from the origin is S, marked with a thick red line.

Table 2.

Meaning of subscripts

L left eye
R right eye
Δ difference between left and right quantities, e.g. vergence angle Hδ = HR−HL
δ half-difference between left and right quantities, e.g. half-vergence HΔ = (HR−HL)/2
c cyclopean eye (mean of left and right), e.g. cyclopean gaze angle Hc = (HR+HL)/2

As noted, VΔ is usually zero. It is also observed that for a given elevation, gaze azimuth and convergence, the torsion of each eye takes on a unique value, which is small and proportional to elevation (Tweed, 1997c). Thus, out of the 6 degrees of freedom, it is a reasonable approximation to consider that the visual system uses only 3: Hc, Vc and HΔ, with VΔ=0, and cyclotorsion Tc and cyclovergence TΔ given by functions of Hc, Vc and HΔ,. Most treatments of physiological vertical disparity have assumed that there is no vertical vergence misalignment or torsion: VΔ=TΔ=Tc=0. We too shall use this assumption in subsequent sections, but we start by deriving the most general expressions that we can. The expressions given in Table 7 assume that all three vergence angles are small, but not necessarily zero. This enables the reader to substitute in realistic values for the cyclovergence TΔ at different elevations (Minken & Van Gisbergen, 1994; Somani, DeSouza, Tweed, & Vilis, 1998; Van Rijn & Van den Berg, 1993), and will later enable us calculate which patterns of vertical disparity are produced by different eye misalignments. The expressions in Table 7 also assume that the interocular distance is small compared to the distance to the viewed object. If the eyes are fixating near the object P, then the small vergence approximation already implies this small baseline approximation, since if P is far compared to the interocular separation, then both eyes need to take up nearly the same posture in order to view it. We believe that this paper is the first to present explicit expressions for two-dimensional retinal disparity that are valid all over the visual field and which do not assume that vertical vergence misalignment, cyclotorsion and cyclovergence are zero.

Under these assumptions, the vertical disparity expressed as the difference in elevation-longitude coordinates is

ηΔcos2ηc(sinTccosHcsinHctanηc)IS+(tanαcsinηccosηccosTcsinTc)HΔ(cosHccosTc+tanαcsinηccosηccosHcsinTc+tanαccos2ηcsinHc)VΔ(tanαccos2ηc)TΔ Equation 1

assuming that I/S, HΔ, TΔ and VΔ are all small.

while the vertical disparity expressed as the difference in elevation-latitude coordinates is

κΔ(cosTccosHcsinαcsinκcsinHccosαcsinκc+sinTccosHccosκc)IRsinTccosαcHΔ(cosHccosTccosαc+sinαc+sinHc)VΔsinαcTΔ Equation 2

assuming that I/R, HΔ, TΔ and VΔ are all small.

The coordinates (αcc) represent the visual direction of the viewed object P in the azimuth-longitude/elevation-longitude coordinate system shown in Figure 3AC, while (αcc) represent visual direction in the azimuth-longitude/elevation-latitude system of Figure 3BD. (αcc) or (αcc) specify the position of P's image on an imaginary cyclopean retina midway between the two real eyes, with gaze azimuth Hc, elevation Vc and torsion Tc.

S and R both represent the distance to the viewed object P. R is the distance of P from the cyclopean point midway between the eyes. S is the length of the component along the cyclopean gaze (Figure 5). These are simply related:

S=Rcosαccosκc=Rcosβccosηc.

As noted, Equation 1 and Equation 2 assume that I/S, I/R, HΔ, VΔ, and TΔ are all small, and they are correct to first order in these terms. However, they make no assumptions about αc, ηc, κc, Hc, Vc and Tc. They are thus valid over the entire retina, not just near the fovea, and for all cyclopean eye positions. Under this small-vergence approximation, the total vertical disparity is the sum of four terms, respectively proportional to one of four possible sources of disparity: (i) the interocular separation as a fraction of object distance, I/R or I/S, (ii) the horizontal vergence HΔ, (iii) vertical vergence error VΔ and (iv) cyclovergence TΔ. Each source of disparity is multiplied by a term that depends on one or both of the components of visual direction (αc and ηc or κc), the gaze azimuth Hc and the overall torsion Tc. So for example, cyclovergence TΔ is multiplied by αc, and so makes no contribution to vertical disparity on the vertical retinal meridian. None of the four disparity terms explicitly depends on elevation Vc, although elevation would affect the disparity indirectly, because it determines the torsion according to Donders' law (Somani, et al., 1998; Tweed, 1997a).

The contribution of the three vergence sources of disparity is independent of object distance; object distance only enters through the interocular separation term (i). This may surprise some readers used to treatments that are only valid near fixation. In such treatments, it is usual to assume that the eyes are fixating the object P, so the vergence HΔ is itself a function of object distance R. We shall make this assumption ourselves in section 3.2 below. However, this section does not assume that the eyes are fixating the object P, so the three vergence angles HΔ, VΔ, and TΔ are completely independent of the object's distance R. Thus, object distance affects disparity only through the explicit dependence on R (or S) in the first term, (i). The contribution of the three vergence terms (ii-iv) is independent of object distance, provided that the visual direction and eye posture is held constant. That is, if we move the object away, but also increase its distance from the gaze axis such that the object continues to fall at the same point on the cyclopean retina, then the contribution of the three vergence terms to the disparity at that point are unchanged. (If the vergence changed to follow the object as it moves away, then of course this contribution would change.).

Of course, expressing disparity as a sum of independent contributions from four sources is valid only to first order. A higher-order analysis would include interactions between the different types of vergence, between vergence and interocular separation and so on. Nevertheless, we and others (e.g.(Garding, et al., 1995) have found that first-order terms are surprisingly accurate, partly because several second-order terms vanish. We believe that the present analysis of disparity as arising from 4 independent sources (interocular separation plus 3 vergence angles) is both new and, we hope, helpful. In the next section, we show how we can use this new analysis to derive expressions for the average vertical disparity experienced in different parts of the visual field.

2. Average vertical disparity expected at different positions on the retina

There have been a few previous attempts to derive the distribution of vertical disparity encountered during normal viewing (Hibbard, 2007; Y. Liu, et al., 2008; Read & Cumming, 2004). However, these studies averaged results across all visual directions. For example, Read & Cumming calculated the distribution of physically possible disparities for all objects whose images fall within 15° of the fovea in both retinas. Critically, they averaged this distribution not only over all possible objects, but over the whole 15° parafoveal area. The spread of their distribution thus reflects both variation in the vertical disparities that are possible at different positions on the retina, and variation that is possible at a single retinal location. To make the distinction clear with a toy example, suppose that all eye position parameters are frozen (Hc=Tc=TΔ=VΔ=0), except for vergence, HΔ, which varies between 0 and 40°, so that elevation-longitude disparity is ηΔ ≈ 0.5HΔtan(αc)sin(2ηc). Under these circumstances, 10° to the left and 10° up from the fovea, vertical disparity would always be positive, running from 0° to +0.9°. On the opposite side of the retina, 10° right and 10° up, vertical disparity would always be negative, running from 0° to −0.9°. Along the retinal meridians, the vertical disparity would always be zero. Read & Cumming's analysis would lump all these together to report that the range of possible vertical disparity is from −0.9° to +0.9°. In other words, the results of Read & Cumming (2004), like those of Hibbard (2007) & Liu et al. (2008), confound variation in the vertical disparity that is possible at a given retinal location with variation across different locations. Similarly, physiological studies that have investigated tuning to vertical disparity have not reported where in the visual field individual neurons were, making it impossible to relate the tuning of these neurons to ecological statistics. For example, one would expect the range of vertical disparity tuning to be narrower for neurons located directly above or below the fovea than for neurons to the “north-west”. The published physiological literature does not make it possible to examine this prediction.

Deriving a full probability density function for ecological disparity requires making assumptions about the eye postures adopted during natural viewing, the scenes viewed and the fixations chosen within each scene. Although there have been major steps towards realistic estimates of these quantities (Hibbard, 2007; Y. Liu, et al., 2008), there is as yet no definitive study, and the issue is beyond the scope of the present paper. However, the expressions derived in the previous section do enable us to estimate the mean vertical disparity as a function of position in the visual field. The previous studies (Hibbard, 2007; Y. Liu, et al., 2008; Read & Cumming, 2004) say only that the mean vertical disparity, averaged across the visual field, is zero; they do not discuss how the mean varies as a function of position. In this section, making only some fairly limited and plausible assumptions about eye position, we shall obtain expressions showing how mean vertical disparity varies as a function of position in the visual field.

The last term in both Equation 1 and Equation 2 depends on eye position only through TΔ, the cyclovergence. According to the extended, binocular versions of Listing's law (Minken & Van Gisbergen, 1994; Mok, Ro, Cadera, Crawford, & Vilis, 1992; Somani, et al., 1998; Tweed, 1997c), this term depends on elevation, and more weakly on convergence. The eyes twist inwards (i.e. top of each eye moves inwards) on looking down from primary position and outwards on looking up, and this tendency is stronger when the eyes are converged: TΔ=Vc(Λ+MHΔ), where Λ and M are constants<1 (Somani, et al., 1998). However, if we assume that people are as likely to look up from primary position as down, and that there is no systematic relationship between elevation and convergence, then we can reasonably approximate the mean cyclovergence as being close to zero: <​TΔ>=0. Thus, the last term in the expressions for both kinds of vertical disparity vanishes.

The next-to-last term is proportional to vertical vergence error, VΔ. Again, we assume that this is on average zero and independent of gaze azimuth or torsion, so that terms like VΔcosHccosTc all average to zero. This assumption may not be precisely correct, but vertical vergence errors are likely to be so small in any case that neglecting this term is not likely to produce significant errors in our estimate of mean vertical disparity.

The next term is proportional to convergence angle HΔ. This is certainly not zero on average. However, part of its contribution depends on sin(Tc), the sine of the cyclopean torsion. Empirically, this is approximately Tc~VHc/2 (Somani, et al., 1998; Tweed, 1997b). So, although cyclotorsion can be large at eccentric gaze angles, provided we assume that gaze is symmetrically distributed about primary position, then <​VHc>=0 and so the mean torsion is zero. Again, in the absence of a particular asymmetry, e.g. that people are more likely to look up and left while converging and more likely to look up and right while fixating infinity, we can reasonably assume that <​HΔsinTc>=0. Equation 1 also contains a term in HΔcosTc. This does not average to zero, but under the assumption that convergence and cyclopean torsion are independent, its mean value is <​HΔ>. Thus, under the above assumptions, Equation 1 and Equation 2 become

ηΔcos2ηc(sinTccosHcsinHctanηc)IS+(tanαcsinηccosηc)HΔ Equation 3
κΔ(cosTccosHcsinαcsinκcsinHccosαcsinκc+sinTccosHccosκc)IR Equation 4

Looking at Equation 3, we see that the average elevation-longitude disparity encountered in natural viewing contains terms like <​sin(Hc)/S>. We now make the reasonable assumption that, averaged across all visual experience, gaze azimuth is independent of distance to the surface. This assumes that there are no azimuthal asymmetries such that nearer surfaces are systematically more likely to be encountered when one looks left, for example. Under this assumption, the term <​sin(Hc)/S> averages to zero. Similarly we assume that <​sin(Tc)cos(Hc)/S>=0. Thus, the entire term in I/S averages to zero. The vast array of different object distances encountered in normal viewing makes no contribution to the mean elevation-longitude disparity at a particular place on the retina. The mean elevation-longitude disparity encountered at position (αcc) is simply

ηΔHΔtanαcsinηccosηc Equation 5

We have made no assumptions about the mean convergence, <​HΔ>, but simply left it as an unknown. It does not affect the pattern of expected vertical disparity across the retina, but merely scales the size of vertical disparities. Convergence is the only eye-position parameter that we cannot reasonably assume is zero on average, and thus it is the only one contributing to mean vertical disparity measured in elevation-longitude.

For elevation-latitude disparity, the dependence on object distance S does not immediately average out. Again, we assume that the terms <​sin(Hc)/R> and <​sin(Tc)cos(Hc)/R> are zero, but this still leaves us with

κΔIRcosHccosTcsinαcsinκc Equation 6

To progress, we need to make some additional assumptions about scene structure. We do this by introducing the fractional distance from fixation, δ:

R=R0(1+δ)

where R0 is the radial distance from the origin to the fixation point (or to the point where the optic axes most nearly intersect, if there is a small vertical vergence error). This is

R0IcosHcHΔ

Thus,

IRcosHcHΔ(1δ) Equation 7

Substituting Equation 7 into Equation 6, we obtain

κΔHΔ(1δ)cosTcsinαcsinκc

We can very plausibly assume that torsion is independent of convergence and scene structure and on average zero, so that <​HΔ(1-δ)cosTc> averages to <​HΔ(1-δ)>. In natural viewing, the distributions of HΔ and d will not be independent (cf. Fig 6 of Liu et al. (2008)). For example, when HΔ is zero, its smallest value, the fixation distance is infinity, and so δ must be negative or zero. Conversely when the eyes are converged on a nearby object (large HΔ), perhaps most objects in the scene are usually further away than the fixated object, making δ predominantly positive. In the absence of accurate data, we assume that the average <​HΔδ> is close to zero. We then obtain

κΔHΔsinαcsinκc Equation 8

Figure 6.

Figure 6

Expected vertical disparity in natural viewing, as a function of position in the cyclopean retina, for (a) elevation-longitude and (b) elevation-latitude definitions of vertical disparity. Vertical disparity is measured in units of <​HΔ>, the mean convergence angle. Because the vertical disparity is small over much of the retina, we have scaled the pseudocolor as indicated in the color-bar, so as to concentrate most of its dynamic range on small values. Contour lines show values in 0.1 steps from −1 to 1.

Equation 8 gives the expected elevation-latitude disparity <​κΔ> as a function of cyclopean elevation-latitude κc, whereas Equation 5 gave the expected elevation-longitude disparity <​ηΔ> as a function of cyclopean elevation-longitude ηc. To make it easier to compare the two, we now rewrite Equation 5 to give the expected elevation-longitude disparity <​ηΔ> as a function of cyclopean elevation-latitude κc. The expected vertical disparity at (αcc) is thus, in the two definitions,

ηΔ(αc,κc)HΔsinαctanκc(cos2αc+tan2κc)κΔ(αc,κc)HΔsinαcsinκc Equation 9

These expressions will strike many readers as familiar. They are the longitude and latitude vertical disparity fields that would be obtained when the eyes adopt their average position, i.e. fixating on the midsagittal plane with no cyclovergence, cyclotorsion or vertical vergence and with the average convergence <​HΔ>, and view a spherical surface centered on the cyclopean point and passing through fixation. The much more general expressions we have considered reduce to this, because vertical-disparity contributions from eccentric gaze, from the fact that objects may be nearer or further than fixation, from cyclovergence and from vertical vergence all cancel out on average. Thus, they do not affect the average vertical disparity encountered at different points in the visual field (although of course they will affect the range of vertical disparities encountered at each position).

It is often stated that vertical disparity increases as a function of retinal eccentricity. Thus, it may be helpful to give here expressions for retinal eccentricity ξ:

cosξ=cosαccosκctan2ξ=tan2αc+tan2ηc Equation 10

Here, eccentricity ξ is defined as the angle PCF, where P is the point on the retina whose retinal eccentricity is being calculated, C is the centre of the eyeball, and F is the fovea.

The mean convergence, <​HΔ>, is not known, but must be positive. It does not affect the pattern of vertical disparity expected at different points in the visual field, but simply scales it. Figure 6 shows the pattern expected for both types of vertical disparity. In our definition, αc and κc represent position on the cyclopean retina, and their signs are thus inverted with respect to the visual field (bottom of the retina represents upper visual field). However, conveniently Equation 9 is unchanged by inverting the sign of both αc and κc, meaning that Figure 6 can be equally well interpreted as the pattern across either the cyclopean retina or the visual field.

Conveniently, within the central 45° or so, the expected vertical disparity is almost identical for the two definitions of retinal vertical disparity we are considering. Near the fovea, the value expected for both types of vertical disparity is roughly <​HΔcκc. Throughout the visual field, the sign of vertical disparity depends on the quadrant. Points in the top-right or bottom-left quadrants of the visual field experience predominantly negative vertical disparity in normal viewing, while points in the top-left or bottom-right quadrants experience predominantly positive vertical disparities. Points on the vertical or horizontal meridians experience zero vertical disparity on average, although the range would clearly increase with vertical distance from the fovea. To our knowledge, no physiological studies have yet probed whether the tuning of disparity-sensitive neurons in early visual areas reflects this retinotopic bias.

3. Properties of elevation-longitude and elevation-latitude definitions of vertical disparity, in the absence of torsion or vertical vergence

So far, we have provided general expressions for vertical disparity but, in order to make comparisons with previous literature more straightforward, in this and subsequent sections we make the simplifying assumption that cyclotorsion, cyclovergence and vertical vergence are all zero (Tc=TΔ=VΔ=0) and demonstrate the consequences of this assumption on the properties of the two types of vertical disparity. In this case, the only degrees of freedom which affect disparity are the horizontal rotation of each eye, expressed as the convergence HΔ and the gaze angle Hc. From Equation 1, elevation-longitude vertical disparity in the absence of torsion and vertical vergence is

ηΔ12sin2ηc(ISsinHc+HΔtanαc) Equation 11

while from Equation 2, elevation-latitude vertical disparity is

κΔIRsinκcsin(αcHc) Equation 12

In general, these two types of vertical disparity have completely different properties. We see that elevation-longitude vertical disparity is zero for all objects, irrespective of their position in space, if the eyes are in primary position, i.e. Hc=HΔ=0. Elevation-latitude vertical disparity is not in general zero when the eyes are in primary position, except for objects on the midline or at infinite distance. Rotating the eyes into primary position does not affect elevation-latitude disparity because, as noted in the Introduction, horizontal rotations of the eyes cannot alter which line of elevation-latitude each point in space projects to; they can only alter the azimuthal position to which it projects. Thus, κΔ is independent of convergence, while gaze azimuth simply sweeps the vertical disparity pattern across the retina, keeping it constant in space. Thus, κΔ depends on gaze azimuth Hc only through the difference (αc-Hc), representing azimuthal position in head-centric space. As a consequence, elevation-latitude vertical disparity is zero for all objects on the midline (X=0, meaning that αc=Hc). The elevation-latitude disparity κΔ at each point in the cyclopean retina is scaled by the reciprocal of the distance to the viewed object at that point.In contrast, elevation-longitude disparity, ηΔ, is independent of object distance when fixation is on the mid-sagittal plane; it is then proportional to convergence. In Table 1, we summarise the different properties of elevation-longitude and elevation-latitude vertical disparity, under the conditions Tc=TΔ=VΔ=0 to which we are restricting ourselves in this section.

Table 1.

Summary of the different properties of the two definitions of retinal vertical disparity, in the absence of vertical vergence error and torsion.

Vertical disparity
defined as:
Properties in the absence of vertical vergence error and torsion (Tc=TΔ=VΔ=0)

Difference in
retinal elevation
longitude, ηΔ
Is zero for objects in plane of gaze.
Is zero when the eyes are in primary position, for objects at any distance anywhere on the
retina.
Increases as eyes converge.
May be non-zero even for objects at infinity, if the eyes are converged.
Is proportional to sine of twice the elevation-longitude.
Is not necessarily zero for objects on the midline.
For fixation on midline, is independent of object distance for a given convergence angle.

Difference in
retinal elevation
latitude, κΔ
Is zero for objects in plane of gaze.
Is zero for objects at infinity.
Is inversely proportional to object's distance.
Is independent of convergence for objects at a given distance.
May be non-zero even when eyes are in primary position.
Is proportional to sine of elevation-latitude.
Is zero for objects on the mid-sagittal plane.

We have seen that ηΔ and κΔ depend very differently on convergence and object distance. For midline fixation, ηΔ is proportional to convergence and is independent of object distance, whereas κΔ is independent of convergence and is inversely proportional to object distance. However, if we consider only objects close to fixation, then object distance and convergence convey the same information. Under these circumstances the two definitions of vertical disparity become similar. This is shown in Figure 7, which plots the vertical disparity field for a frontoparallel plane. The two left panels show elevation-longitude vertical disparity; the two right panels show elevation-latitude vertical disparity. In the top row, the eyes are viewing a frontoparallel plane at a distance 60cm, and in the bottom row, a plane at 10m. In each case, the eye position is the same: looking 15° to the left, and converged so as to fixate the plane at 60cm.

Figure 7.

Figure 7

Vertical disparity field all over the retina, where the visual scene is a frontoparallel plane, i.e. constant head-centered coordinate Z. AB: Z=60cm; CD: Z=10m. The interocular distance was 6.4cm, gaze angle Hc=15° and convergence angle HΔ=5.7°, i.e. such as to fixate the plane at Z=60cm. Vertical disparity is defined as difference in (AC) elevation-longitude, (BD) elevation-latitude. Lines of azimuth-longitude and (AC) elevation-longitude, (BD) elevation-latitude are marked in black in 15° intervals. The white line shows where the vertical disparity is zero. The fovea is marked with a black dot. The same pseudocolor scale is used for all four panels. Note that the elevation-longitude disparity, ηΔ, goes beyond the colorscale at the edges of the retina, since it tends to infinity as |αc| tends to 90°. This figure was generated by DiagramOfVerticalDisparity_planes.m in the Supplementary Material.

In Figure 7AB, where the eyes are fixating the viewed surface, both definitions of vertical disparity give similar results, especially near the fovea. For both definitions, vertical disparity is zero for objects in the plane of gaze (Y=0, i.e. ηcc=0), and also along a vertical line whose position depends on the gaze angle. For elevation-latitude disparity κΔ, this line is simply the line of azimuth-longitude αc=Hc, here 15°. This is the retinal projection of the mid-sagittal plane, X=0. That is, in the absence of torsion or vertical vergence error, elevation-latitude vertical disparity is zero for objects on the midline, independent of their distance or of the convergence angle. For elevation-longitude vertical disparity ηΔ, no such simple result holds. The locus of zero vertical disparity (vertical white line in Figure 7AC) depends on object distance and the eyes' convergence, as well as gaze angle. However, for objects relatively near fixation, these differences are minor, so the locus of zero ηΔ is also close to 15°.

It is not always the case, however, that the differences between elevation-longitude and elevation-latitude vertical disparities are minor. Figure 7CD shows the two vertical disparity fields for a surface at a much greater distance from the observer than the fixation point. The position of the eyes is the same as in AB, but now the viewed surface is a plane at 10m from the observer. Now, the pattern of vertical disparities is very different. As we saw from Equation 12, elevation-latitude vertical disparity is zero for all objects at infinity, no matter what the vergence angle HΔ. Thus, for Z=10m, it is already close to zero across the whole retina (Figure 7D). Elevation-longitude vertical disparity does not have this property. It is zero for objects at infinity only if the eyes are also fixating at infinity, i.e. HΔ=0. Figure 7C shows results for HΔ=5.7°, and here the second term in Equation 11 gives non-zero vertical disparity everywhere except along the two retinal meridians.

In summary, the message of this section is that elevation-longitude and elevation-latitude definitions of vertical disparity give very similar results for objects near the fovea and close to the fixation distance.. However, when these conditions are not satisfied, the two definitions of vertical disparity can produce completely different results.

3.1. Epipolar lines: Relationship between vertical and horizontal disparity

Disparity is a two-dimensional quantity, but for a given eye position, not all two-dimensional disparities are physically possible. Figure 8A shows how the physically possible matches for the red dot on the left retina fall along a line in the right retina. The object projecting to the red dot could lie anywhere along the red line shown extending to infinity, and each possible position implies a different projection onto the right retina. The set of all possible projections is known in the literature as an epipolar line (Hartley & Zisserman, 2000).

Figure 8.

Figure 8

Epipolar line and how it differs from the “line of possible disparities” shown in Figure 9. A: How an epipolar line is calculated: it is the set of all possible point on the right retina (heavy blue curve) which could correspond to the same point in space as a given point on the left retina (red dot). B: Epipolar line plotted on the planar retina. Blue dots show 3 possible matches in the right eye for a fixed point in the left retina (red dot). The cyclopean location or visual direction (mean of left and right retinal positions, black dots) changes as one moves along the epipolar line. C: Possible matches for a given cyclopean position (black dot). Here, we keep the mean location constant, and consider pairs of left/right retinal locations with the same mean. D: Line of possible disparities implied by the matches in B. These are simply the vectors linking left to right retinal positions for each match (pink lines). Together, these build up a line of possible disparities (green line).

This definition of epipolar line treats the eyes asymmetrically: one considers a point in one eye, and the corresponding line in the other eye. Everywhere else in this paper, we have treated the eyes symmetrically, rewriting left and right eye coordinates in terms of their sum and difference: position on the cyclopean retina, and disparity. So in this section, we shall consider something slightly different from the usual epipolar lines: we shall consider the line of possible disparities at a given point in the cyclopean visual field. Figure 8BCD shows how this differs from an epipolar line. As one moves along an epipolar line (Figure 8B), not only the two-dimensional disparity, but also the cyclopean position, varies. We shall consider how disparity varies while keeping cyclopean position constant (Figure 8D).

To achieve this, we need to express vertical disparity as a function of horizontal disparity. So far in this paper, we have expressed vertical disparity as a function of object distance and eye position. Of course, horizontal disparity is also a function of object distance and eye position. So if we substitute in for object distance using horizontal disparity, we obtain the function relating horizontal and vertical disparity, for a given eye position and location in the visual field. Using the expressions given in the Appendix, it is simple to obtain the line of possible disparities for arbitrary cyclovergence, cyclotorsion and vertical vergence error. In this section, for simplicity, we continue to restrict ourselves to Tc=TΔ=VΔ=0. Under these circumstances, azimuth-longitude horizontal disparity is (Appendix B:, Table 10)

αΔIScosαccos(αcHc)+HΔ Equation 13
Table 10.

Expressions for horizontal disparity in different coordinate systems. These are correct to first-order in interocular distance I/S (I/R) and in the convergence angle HΔ. They assume cyclotorsion, cyclovergence and vertical vergence are all zero: Tc=TΔ=VΔ=0. They hold all over the retina and for any cyclopean gaze Hc or elevation Vc.

Horizontal disparity For zero torsion and vertical vergence error
In planar Cartesian retinal coordinates as a function of retinal
location in planar Cartesian coordinates:
xΔ(xcHcsin+Hccos)IS+(xc2+1)HΔ
In planar Cartesian retinal coordinates as a function of spatial
position in gaze-centered coordinates
xΔIS2(UHcsinSHccos)+U2+S2S2HΔ
In azimuth-longitude, as a function of spatial location in
gaze-centered coordinates
αΔIS2+U2(UHcsinSHccos)+HΔ
In azimuth-longitude, as a function of retinal location in
azimuth-longitude/elevation-longitude coordinates
αΔIScosαccos(αcHc)+HΔ
In azimuth-longitude, as a function of retinal location in
azimuth-longitude/elevation-latitude coordinates
(same as above since αΔ is then independent of retinal elevation).
In azimuth-latitude, as a function of spatial location in
gaze-centered coordinates
βΔ=I(HcsinUSHccos(W2+S2))(U2+W2+S2)W2+S2+SHΔW2+S2
In azimuth-latitude, as a function of retinal location in
azimuth-latitude/elevation-longitude coordinates
βΔ=IScosβccosηc(Hcsinsinβccosηc+Hccoscosβc)+cosηcHΔ

If we use horizontal disparity to substitute for object distance in Equation 11, we obtain the following relationship between horizontal (azimuth-longitude) and vertical (elevation-longitude) disparity:

ηΔ12sin2ηcsecαc(HΔsinαc(αΔHΔ)sec(αcHc)sinHc)

For elevation-latitude vertical disparity, again substituting for object distance in Equation 12, we obtain

κΔ12(sin2κc)(αΔHΔ)tan(Hcαc)

Thus, the vertical disparities that are geometrically possible at a given position in the visual field are a linear function of the horizontal disparity. This is shown for elevation-latitude disparity by the green line in Figure 9. Where we are on this line depends on object distance.

Figure 9.

Figure 9

The thick green line shows the line of two-dimensional disparities that are physically possible for real objects, for the given eye posture (specified by convergence HΔ and gaze azimuth Hc) and the given visual direction (specified by retinal azimuth αc and elevation κc). The green dot shows where the line terminates on the abscissa. For any given object, where its disparity falls on the green line depends on the distance to the object at this visual direction. The white circle shows one possible distance. Although, for clarity, the green line is shown as having quite a steep gradient, in reality it is very shallow close to the fovea. Thus, it is often a reasonable approximation to assume that the line is flat in the vicinity of the distance one is considering (usually the fixation distance), as indicated by the horizontal green dashed line. This is considered in more detail in the next section.

Note that this expression is not valid for if either αc or (αc-Hc) is 90°, since then horizontal disparity is independent of object distance (Equation 13). So for example if we are considering an azimuthal direction of 45° (α=45°) and the eyes are looking off 45° to the right (Hc=−45°), this expression fails. Apart from this relatively extreme situation, it is generally valid.

Note also that the line of possible disparities does not extend across the whole plane of disparities. We have adopted a sign convention in which “far” disparities are positive. The largest possible horizontal disparity occurs for objects at infinity. Then, we see from Equation 13 that the horizontal disparity is equal to the convergence angle, HΔ. For objects closer than infinity, the horizontal disparity is smaller, becoming negative for objects nearer than the fixation point. Thus, the green line in Figure 9 terminates at αΔ=HΔ. The elevation-latitude vertical disparity at this point in the visual field thus has only one possible sign, either negative or positive depending on the sign of κc(Hcc) (since (αΔ-HΔ) is always negative). For elevation-latitude vertical disparity, the eye-position parameters have a particularly simple effect on the line of possible disparities. The convergence angle HΔ controls the intercept on the abscissa, i.e. the horizontal disparity for which the vertical disparity is zero. The gradient of the line is independent of convergence, depending only on the gaze angle. To avoid any confusion, we emphasise that this “disparity gradient” is the rate at which vertical disparity would change if an object slid nearer or further along a particular visual direction, so that its horizontal disparity varied while its position in the (cyclopean) visual field remained constant. Thus, we are considering the set of two-dimensional disparities that can be produced by a real object for a given binocular eye position. This might theoretically be used by the visual system in solving the stereo correspondence problem if eye position were known. This “disparity gradient” is not the same as the disparity derivatives discussed below, Section 4, in the context of deriving eye position given a solution of the correspondence problem, which concern the rate at which vertical disparity changes as a function of visual direction in a given scene.

In Figure 9, the gradient of the green line is exaggerated for clarity. In fact, when κc and (Hcc) are both small (i.e. for objects near the midline and near the plane of regard), the gradient is close to zero. Even quite large changes in horizontal disparity produce very little effect on vertical disparity. In these circumstances, it is reasonable to approximate vertical disparity by its value at the chosen distance, ignoring the gradient entirely. We go through this in the next section.

3.2. For objects near fixation, vertical disparity is independent of object distance

It is often stated that, to first order, vertical disparity is independent of object distance, depending only on eye position (e.g. (Garding, et al., 1995; Read & Cumming, 2006). Horizontal disparity, in contrast, depends both on eye position and object distance. Thus, vertical disparity can be used to extract an estimate of eye position which can then be used to interpret horizontal disparity.

At first sight, these statements appear to conflict with much of this paper. Consider the first-order expressions for vertical disparity given in Equation 11 and Equation 12. Both depend explicitly on the object distance (measured radially from the origin, R, or along the gaze direction, S, Figure 5). Figure 7AB vs Figure 7CD, which differ only in the object distance, show how both types of vertical disparity depend on this value. Elevation-latitude disparity does not even depend on the convergence angle HΔ, making it appear impossible to reconstruct vergence from measurements of elevation-latitude disparity alone.

This apparent contradiction arises because the authors quoted were considering the disparity of objects near to fixation. Our Equation 11 and Equation 12, in contrast, are valid for all object locations, provided only the object distance is large compared to the interocular distance (small baseline approximation). We now restrict ourselves to the vicinity of fixation. That is, we assume that the object is at roughly the same distance as the fixation point. We express the radial distance to the object, R, as a fraction of the distance to fixation, R0:

R=R0(1+δ) Equation 14

Under our small vergence angle approximation, the radial distance to the fixation point is

R0IcosHcHΔ Equation 15

For small values of d, then, we have

IRHΔ(1δ)cosHcISHΔ(1δ)cosαccosκccosHc Equation 16

Note that this breaks down at Hc =90°. This is the case where the eyes are both directed along the interocular axis. Then, the distance to the fixation point is undefined, and we cannot express R as a fraction of it. The case Hc =90° is relevant to optic flow, but not to stereo vision. Our analysis holds for all gaze angles that are relevant to stereopsis.

From Equation 11 and Equation 12, using the fact that R=Ssecαcsecκc, the two definitions of vertical disparity then become

ηΔsinηccosηc(HΔ(1δ)cosHcsecαcsecκcsinHc+HΔtanαc)κΔHΔ(1δ)cosHcsinκcsin(αcHc) Equation 17

The dependence on object distance is contained in the term δ, the fractional distance from fixation. But by assumption, this is much smaller than 1. The vertical disparities are dominated by terms independent of distance; to an excellent approximation, we have

ηΔHΔsinηccosηc(tanαctanHcsecαc1+tan2ηccos2αc)κΔHΔcosHcsinκcsin(αcHc) Equation 18

where we have used tanκ=tanηcosα to substitute for elevation-latitude κ in the expression for elevation-longitude vertical disparity, ηΔ.

Thus, for objects at the same distance as fixation, the dependence on object distance can be expressed as the convergence angle. Changes in scene depth produce negligible changes in vertical disparity: to a good approximation, vertical disparity is independent of scene structure, changing only with slow gradients which reflect the current binocular eye position. This statement is true all across the retina (i.e. for all αc and κc).

For horizontal disparity, the situation is more subtle. It can be shown that, under the same approximations used in Equation 17, azimuth-longitude horizontal disparity is given by

αΔHΔ(1δ)cosHcsecκccos(αcHc)+HΔ. Equation 19

This equation resembles Equation 17, so at first sight it seems that we can drop d as being small in comparison with 1, meaning that horizontal disparity is also independent of object distance. If Hc, αc and κc are large, this is correct. Under these “extreme” conditions (far from the fovea, large gaze angles), horizontal disparity behaves just like vertical disparity. It is dominated by eye position and location in the visual field, with object distance making only a small contribution. However, the conditions of most relevance to stereo vision are those within ~10° of the fovea, where spatial resolution and stereoacuity is high. In this region, a key difference now emerges between horizontal and vertical disparity: Vertical disparity becomes independent of scene structure, whereas horizontal disparity does not. The terms in Equation 19 that are independent of object distance δ cancel out nearly exactly, meaning that the term of order δ is the only one left. Thus, horizontal disparity becomes

αΔHΔ(δαctanHc)(parafoveal approximation)

This expression is valid near the fixation point (δ, αc, κc all small) and for gaze angles that do not approach 90° (where Equation 19 diverges). Near the fovea, elevation-latitude and elevation-longitude become indistinguishable (see lines of latitude and longitude in Figure 7). For the near-fixation objects we are considering, therefore, elevation-latitude and elevation-longitude definitions of vertical disparity we derived previously (Equation 18) become identical and both equal to

ηΔκΔHΔκc(αctanHc)(parafoveal approximation) Equation 20

Critically, this means that for the near-fixation case most relevant to stereo vision, horizontal disparity reflects scene structure as well as eye position, whereas vertical disparity depends only on eye position. This means that estimates of eye position, up to elevation, can be obtained from vertical disparity and used to interpret horizontal disparity. In this section, we have set 3 of the 6 eye position parameters – cyclovergence, cyclotorsion and vertical vergence error – to zero, meaning that we only have 2 eye position parameters left to extract. Thus before proceeding, we shall generalize, in Section 4, to allow non-zero values for all 6 eye position parameters. We shall then show how 5 of these parameters can be simply derived from the vertical disparity in the vicinity of the fovea.

4. Obtaining eye position from vertical disparity and its derivatives at the fovea

In this section, we derive approximate expressions for 5 binocular eye position parameters in terms of the vertical disparity and its derivatives near the fovea. As throughout this paper, we work in terms of retinal disparity, since this is all that is available to the visual system before eye position has been computed. We do not require any special properties of the viewed surface other than that it is near fixation and smooth, so that all derivatives exist. We allow small amounts of cyclotorsion, cyclovergence and vertical vergence error, but restrict ourselves to small gaze angles. Mathematically, this means we approximate cosHc~1 and sinHc~Hc. In Figure 11, we show that our results hold up well at least out to Hc=15°. This is likely to cover most gaze angles adopted during natural viewing. We work in the vicinity of the fovea, so retinal azimuth αc and elevation κc are also both small. In this case, the distinction between latitude and longitude becomes immaterial. We shall write our expressions in terms of elevation-latitude κ, but in this foveal approximation, the same expressions would also hold for elevation-longitude η. We shall show how our equations for the vertical disparity field, κΔ, can be used to read off gaze angle, convergence, cyclovergence, cyclotorsion and vertical vergence.

Figure 11.

Figure 11

Scatterplots of estimated eye position parameters against actual values, both in degrees, for 1000 different simulated eye positions. Black lines show the identity line. Some points with large errors fall outside the range of the plots, but the quoted median absolute errors are for all 1000 simulations. On each simulation run, eye position was estimated as follows. First, the viewed surface was randomly generated. Head-centered X and Y coordinates were generated randomly near the fixation point (XF,YF,ZF). Surface Z-coordinates were generated from Zd = Σij aijXdiYdj, where Xd is the X-position relative to fixation, Xd=X-XF (Yd, Zd similarly, all in cm), i and j both run from 0 to 3, and the coefficients aij are picked from a uniform random distribution between ±0.02 on each simulation run. This yielded a set of points on a randomly-chosen smooth 3D surface near fixation. These points were then projected to the retinas, and the vertical disparity within 0.5° of the fovea was fitted with a parabolic surface. This simulation is Matlab program ExtractEyePosition.m in the Supplementary Material.

We begin with Equation 7, which expressed I/R in terms of horizontal vergence HΔ and the fractional distance of an object relative to fixation, δ. If there is a vertical vergence error VΔ, then there will not be a fixation point, because gaze rays will not intersect. However, Equation 7 is still valid, with δ interpreted as a fraction of the distance to the point where the gaze rays most closely approach each other. We substitute Equation 7 into our most general expression for vertical disparity, Equation 2, and make the additional approximation that the gaze azimuth Hc and overall torsion Tc are both small:

κΔ(sinαcsinκcHccosαcsinκc+Tccosκc)HΔ(1δ)(Tccosαc)HΔ(cosαc+Hcsinαc)Vc(sinαc)TΔ Equation 21

When we finally make the approximation that we are near the fovea, i.e. that αc and κc are also small, we find that the lowest-order terms are

κΔ=VΔαcTΔ+HΔκc(αcHc)δHΔTcαcHcVΔ Equation 22

Because we are allowing non-zero torsion Tc, vertical disparity now also depends on object distance, through δ. However, this is a third-order term. To first-order, the vertical disparity at the fovea measures any vertical vergence error. Thus, we can read off vertical vergence VΔ simply from the vertical disparity measured at the fovea:

VΔ~κΔ

To derive expressions for the remaining eye position parameters, we will need to differentiate Equation 21 with respect to direction in the visual field. We will use subscripts as a concise notation for differentiation: for example, κΔα indicates the first derivative of the vertical disparity κΔ with respect to azimuthal position in the visual field, αc, holding the visual field elevation κc constant. Similarly, κΔακ is the rate at which this gradient itself alters as one moves vertically:

κΔακΔαcκcκΔακκcαcαcκcκΔ

Note that these derivatives examine how vertical disparity changes on the retina as the eyes view a given, static scene. This is not to be confused with the gradient discussed in Figure 9, which considered how vertical disparity varies as an object moves in depth along a particular visual direction. We assume that the visual scene at the fovea consists of a smooth surface that remains close to fixation in the vicinity of the fovea. The surface's shape is specified by δ, the fractional difference between the distance to the surface distance and the distance to fixation. In Section 3.2, we were considering a single point in the cyclopean retina, and so δ was just a number: the fractional distance at that point. Since we are now considering changes across the retina, δ is now a function of retinal location, δ(αcc). The first derivatives of δ specify the surface's slant, its second derivatives specify surface curvature, and so on. δ and its derivatives δa, etc, are assumed to remain small in the vicinity of the fovea.

After performing each differentiation of Equation 21, we then apply the parafoveal approximation and retain only the lowest-order terms. In this way, we obtain the following set of relationships between derivatives of the vertical disparity field and the binocular eye position parameters:

κΔVΔ Equation 23
κΔαTΔ Equation 24
κΔκHcHΔ Equation 25
κΔακHΔ Equation 26
κΔκκTcHΔ Equation 27

assuming that αc, κc, δ, δα, δκ, δαα, δκκ, Hc, Tc, HΔ, Tc, Vc are all small.

To lowest-order, there is no dependence on scene structure: under this near-fixation approximation, vertical disparity and its derivatives depends only on eye position. Each term enables us to read off a different eye-position parameter. Any vertical disparity at the fovea reflects a vertical vergence error (Equation 23 ;(Howard, Allison, & Zacher, 1997; Read & Cumming, 2006)). The rate at which vertical disparity changes as we move horizontally across the visual field, sometimes called the vertical shear disparity (Banks, et al., 2001; Kaneko & Howard, 1997b), tells us the cyclovergence (Equation 24). A “saddle” pattern, i.e. second derivative with respect to both horizontal and vertical disparity, tells us the vergence (Equation 26; (Backus, et al., 1999)). And then the rate at which vertical disparity changes as we move vertically across the visual field tells us the gaze angle (Equation 25; (Backus & Banks, 1999; Banks & Backus, 1998; Gillam & Lawergren, 1983; Mayhew, 1982; Mayhew & Longuet-Higgins, 1982)). Finally, the second derivative provides an estimate of cyclopean cyclotorsion (Equation 27). Although many of these relationships with aspects of eye position have been identified in the past, it is useful to be able to identify the extent to which the approximations hold under a range of eye positions.

The relationships given in Equation 23-Equation 27 provide an intuitive insight into how different features of the parafoveal vertical disparity field inform us about eye position. The approximations used lead to some small errors, but they are sufficiently small to ignore under most circumstances. For example, retaining 3rd-order terms in the expression for the second derivative κΔκκ yields

κΔκκHΔTcHΔδκκTc2HΔδκ(αcHc)+HΔ(Hcαc)κc+δHΔTc.

If torsion Tc is zero, then near the fovea κΔκκ will in fact be dominated by a term depending on the rate of change of δ as we move vertically in the visual field, reflecting surface slant:

κΔκκ2HΔHcδκ. Equation 28

Applying the recipe in Equation 27 would lead us to conclude Tc ≈ −2Hcδκ, instead of the correct value of zero. Now Equation 27 was derived assuming small Hc and δκ, so the misestimate will be small, but nevertheless present. In Figure 11, we examine how well our approximations bear up in practice. Each panel shows the eye position parameters estimated from Equation 23-Equation 27 plotted against their actual values, for 1000 different simulations. On each simulation run, first of all a new binocular eye posture was generated, by picking values of Hc, Tc, Vc, HΔ, TΔ, VΔ randomly from uniform distributions. Torsion Tc, cyclovergence TΔ and vertical vergence error Vc are all likely to remain small in normal viewing, and were accordingly picked from uniform distributions between ±2°. Gaze azimuth and elevation were picked from uniform distributions between ±15°. Convergence was picked uniformly from the range 0-15°, representing viewing distances from infinity to 25cm or so. Note that it is not important, for purposes of testing Equation 23-Equation 27, to represent the actual distribution of eye positions during natural viewing, but simply to span the range of those most commonly adopted. A random set of points in space was then generated in the vicinity of the chosen fixation point. The X and Y coordinates of these points were picked from uniform random distributions, and their Z coordinate was then set according to a function Z(X,Y), whose exact properties were picked randomly on each simulation run, but which always specified a gently curving surface near fixation (for details, see legend to Figure 11). The points were then projected onto the two eyes, using exact projection geometry with no small-baseline or other approximations, and their cyclopean locations and disparities were calculated. In order to estimate derivatives of the local vertical disparity field, the vertical disparities of points within 0.5° of the fovea, of which there were usually 200 or so, were then fitted with a parabolic function:

κΔ=c0+c1αc+c2κc+c3αc2+c4κ2+c5ακ.

The fitted coefficients ci were then used to obtain estimates of vertical disparity and its gradients at the fovea (κΔα=c1, etc). Finally, these were used in Equation 23-Equation 27 to produce the estimates of eye position shown in Figure 11.

The results in Figure 11 show that most eye position parameters can be recovered with remarkable accuracy. The worst is the cyclotorsion Tc, which is recovered quite accurately (say to within 10 arcmin) about half the time, but the rest of the time is widely scattered, for the reasons discussed around Equation 28. Nevertheless, overall performance is good, with a median error of <0.3°. This shows that the our simple, intuitive analytical expressions relating eye position to vertical disparity (Equation 23-Equation 27) are reliable under most circumstances..

This is a theoretical paper, and the expressions above (Equation 23-Equation 27) are simply a mathematical statement, spelling out the relationships that exist between eye position and vertical disparity. Does the visual system, in fact, use retinal eye position estimates extracted from the disparity field? In the case of vertical vergence, cyclotorsion and cyclovergence, the answer seems to be yes. Disparity fields indicating non-zero values of these parameters elicit corrective eye movements tending to null the retinal disparity, suggesting that the retinal disparity field was taken as evidence of ocular misalignment (Carpenter, 1988; Howard, 2002). In the case of gaze azimuth and convergence, the use is more subtle. There is little evidence that the perceived headcentric direction of a stimulus corresponds to the gaze azimuth indicated by its vertical disparity field (Banks, et al., 2002; Berends, et al., 2002). Rather, retinal estimates of gaze azimuth and vergence seem to be used to convert horizontal disparity directly into estimates of surface slant and curvature.

Here, we briefly sketch this process, showing how the eye position parameters obtained from Equation 23-Equation 27 can be used to interpret horizontal disparity, αΔ. In the domain we are considering, horizontal disparity differs from vertical disparity in that it is affected by the viewed scene as well as eye position (recall that Equation 22 showed that, to lowest order, vertical disparity is independent of scene structure). The horizontal disparity itself depends on the distance of the viewed surface relative to fixation, δ, while its first derivatives reflect the surface's slant. Again retaining terms to lowest order, it can be shown that

αΔδHΔTcVΔ+κcTΔαΔαδαHΔHΔ(Hcαc)κcVΔαΔκδκHΔHΔκc+(Hcαc)VΔ+TΔ Equation 29

where δα is the rate of change of δ as we move horizontally in the visual field, δα=∂δ/∂α|κ. It is a measure of surface slant about a vertical axis, and δκ, defined analogously, reflects surface slant about a horizontal axis. δ is a dimensionless quantity, the fractional distance from the fixation point, but the derivative δα is approximately equal to Rα/R, where R is the distance to the fixated surface and Rα is the rate at which this distance changes as a function of visual field azimuth. Thus, δα is the tangent of the angle of slant about a vertical axis, while δκ represents the tangent of the angle of slant about a horizontal axis. We can invert Equation 29 to obtain estimates of surface distance and slant in terms of horizontal disparity and eye position, and then substitute in the eye position parameters estimated from vertical disparity. Note that the estimates of surface slant are unaffected by small amounts of cyclotorsion, Tc. This is convenient for us, since cyclotorsion was the aspect of eye position captured least successfully by our approximate expressions (Figure 11).

From Equation 23-Equation 27 and Equation 29, we can solve for δ and its derivatives in terms of horizontal and vertical disparity and their derivatives:

δαΔ+κcκΔακΔακ+κΔκκκΔκΔακ2δααΔακΔκκcκΔκΔακ Equation 30
δκ(αΔκ+κΔα)κΔακκΔκΔκκΔακ2+κc

These expressions relate horizontal and vertical retinal disparity directly to surface properties, without any explicit dependence on eye position. It seems that the visual system does something similar. As many previous workers have noted, the visual system appears to use purely local estimates, with no attempt to enforce global consistency in the underlying eye postures implicit in these relationships. Thus, values of surface slant consistent with opposite directions of gaze (left vs right) can simultaneously be perceived at different locations in the visual field (Allison, Rogers, & Bradshaw, 2003; Kaneko & Howard, 1997a; Pierce & Howard, 1997; Rogers & Koenderink, 1986; Serrano-Pedraza, Phillipson, & Read, 2009). Enforcing consistency across different points of the visual field would require lateral connections which might be quite costly in cortical wiring, and would be completely pointless, since in the real world eye position must always be constant across the visual field (Adams, et al., 1996; Garding, et al., 1995).

4.1. Relationship to previous literature

There is a substantial literature on obtaining metric information about scene structure from two-dimensional disparity. It can be divided into two fairly distinct categories: “photogrammetric” and “psychological”. The first comes mainly from the computer vision community (Hartley & Zisserman, 2000; Longuet-Higgins, 1981). Here, one uses the two-dimensional disparities of a limited number of point correspondences to solve for binocular eye position, and then back-projects to calculate each object's 3D location in space. The second approach is commoner in the psychological literature (Backus & Banks, 1999; Backus, et al., 1999; Banks, et al., 2002 ; Banks, et al., 2001; Kaneko & Howard, 1996, 1997b; Koenderink & van Doorn, 1976 ; Rogers & Bradshaw, 1993, 1995; Rogers & Cagenello, 1989). Here, one calculates quantities such as horizontal and vertical size ratios, which are effectively local derivatives of disparity, and uses these either to extract estimates of eye position parameters (Banks, et al., 2001; Kaneko & Howard, 1997b) or to move directly to scene properties such as surface slant, without computing an explicit estimate of eye position. These two approaches are closely related (Adams, et al., 1996; Garding, et al., 1995). In the photogrammetric approach, the point correspondences can be anywhere in the visual field (subject to certain restrictions, e.g. not all collinear (Longuet-Higgins, 1981)). If the points all happen to be closely spaced together, then they contain the same information as the derivatives of disparity at that location. Thus, in this regard the psychological literature represents a special case of the photogrammetric approach: extracting eye position from a particular set of correspondences.

However, the photogrammetric literature does not provide explicit expressions for eye position in terms of disparity; rather, eye position is given implicitly, in large matrices which must be inverted numerically. Because the treatment is fully general, the distinction between horizontal and vertical disparity is not useful (e.g., because a torsion of 90° transforms one into the other, or because some epipolar lines become vertical as gaze azimuth approaches 90°). Thus, in the machine vision literature, disparity is considered as a vector quantity, rather than analysed as two separate components. The psychological literature is less general, but offers a more intuitive understanding of how eye position affects disparity in the domain most relevant to natural stereo viewing (objects close to the fovea, eyes close to primary position). As we saw in the previous section, in this domain, disparity decomposes naturally into horizontal and vertical components, which have different properties. Critically, in this domain, vertical disparity is essentially independent of scene structure, and eye position can be estimated from this component alone.

However, as far as we are aware, no paper gives explicit expressions for all eye position parameters in terms of retinal vertical disparity. Much of the psychological literature jumps straight from disparity derivatives to properties such as surface slant, without making explicit the eye position estimates on which these implicitly depend. In addition, the psychological literature can be hard to follow, because it does not always make it clear exactly what definition of disparity is being used. Sometimes, the derivation appears to use optic array disparity, so it is not clear how the brain could proceed given only retinal disparity; or the derivation appears to rely on special properties of the scene (e.g. it considers a vertically-oriented patch), and it is not clear how the derivation would proceed if this property did not hold. Our derivation makes no assumptions about surface orientation, and is couched explicitly in retinal disparity.

Our expression for δα, Equation 30, is a version of the well-known expressions deriving surface slant from horizontal and vertical size ratios (Backus & Banks, 1999; Backus, et al., 1999; Banks, et al., 2002 ; Banks, et al., 2001; Kaneko & Howard, 1996, 1997b; Koenderink & van Doorn, 1976 ; Rogers & Bradshaw, 1993, 1995; Rogers & Cagenello, 1989). “Horizontal size ratio” or HSR is closely related to the rate of change of horizontal disparity as a function of horizontal position in the visual field, whereas “vertical size ratio” reflects the gradient of vertical disparity as a function of vertical position. In the notation of Backus & Banks (1999), for example, which defines HSR and VSR around the fixation point,

ln(HSR)αΔα.ln(VSR)κΔκ.

In their notation, S = surface slant, so our δα = tan(S), and convergence, our HΔ≈κΔακ, is μ. Thus if there is no vertical disparity at the fovea, Equation 30 becomes

δαtanSαΔακΔκκΔακ1μln(HSRVSR) Equation 31

which is Equation 1 of Backus et al (1999) and Backus & Banks (1999).

This relationship has been proposed as an explanation of the induced effect. In the induced effect, one eye's image is stretched vertically by a factor m about the fixation point, thus adding a term m κc to the vertical disparity field. The vertical disparity at the fovea is still zero, and the only vertical disparity derivative to be affected is κΔκ, which gains a term m. This causes a misestimate of surface slant about a vertical axis:

δα,estδα,truemHΔ

Size-ratio-based theories of the induced effect are often contrasted with photogrammetric approach based on misestimates of gaze angle (Clement, 1992; Garding, et al., 1995; Longuet-Higgins, 1982; Mayhew & Longuet-Higgins, 1982; Ogle, 1952). Size-ratio theories use local disparity derivatives to produce a direct estimate of slant. Photogrammetric theories use a set of point correspondences distributed across the whole retina is used to obtain an estimate of eye position, which is then used to interpret horizontal disparity. The treatment here makes clear that the mathematics underlying both theories is really the same. Suppose that in the photogrammetric approach, the point correspondences are close together in the visual field. The points project to slightly different points on the cyclopean retina, and have slightly different disparities. We can express these differences as disparity gradients on the cyclopean retina, or equivalently as size ratios. From these disparity gradients we can derive eye posture, and hence the surface distance and slant. Thus, both size-ratio and photogrammetric explanations of the induced effect rely, mathematically, on the fact that vertical magnification can be interpreted as a misestimate of gaze angle. This is obscured in Equation 31 because there is no explicit mention of gaze angle, but in fact, as we see by comparing Equation 23-Equation 27 and Equation 29, the reason that VSR is useful in interpreting horizontal disparity is because it acts as a proxy for gaze angle (Adams, et al., 1996).

The real difference between the theories is the scale at which they operate (Adams, et al., 1996; Garding, et al., 1995). Mayhew and Longuet-Higgins originally described an algorithm for fitting a unique eye posture to the correspondences across the whole retina (Mayhew, 1982; Mayhew & Longuet-Higgins, 1982). This would seem to be the best strategy for obtaining the most reliable estimate of eye position, and for that reason is the approach used in computer vision. But as noted in Section 4, the brain appears to proceed locally, at least in the case of surface slant. That is, it directly estimates surface slant from local disparity derivatives, as in Equation 30, without checking that the eye postures implied by these local derivatives are globally consistent. There seems to be considerable inter-subject variation in what “local” means, ranging from as large as 30° for some subjects down to 3° for others (Kaneko & Howard, 1997a; Serrano-Pedraza, et al., 2009).

Discussion

Vertical disparity has been much discussed in recent years (Adams, et al., 1996; Backus & Banks, 1999; Backus, et al., 1999; Banks & Backus, 1998; Banks, et al., 2002; Banks, et al., 2001; Berends & Erkelens, 2001; Berends, et al., 2002; Bishop, 1989; Brenner, et al., 2001; Clement, 1992; Cumming, Johnston, & Parker, 1991; Garding, et al., 1995; Gillam, Chambers, & Lawergren, 1988; Kaneko & Howard, 1997a; Longuet-Higgins, 1981, 1982; Mayhew, 1982; Mayhew & Longuet-Higgins, 1982; Read & Cumming, 2006; Rogers & Bradshaw, 1993; Schreiber, Crawford, Fetter, & Tweed, 2001; Serrano-Pedraza, et al., 2009; Serrano-Pedraza & Read, 2009; Stenton, Frisby, & Mayhew, 1984; Stevenson & Schor, 1997). However, progress has been hampered by the lack of a clear, agreed set of definitions. In the Introduction, we identified no fewer than 4 definitions of vertical disparity: two types of optic-array disparity, and two types of retinal disparity. Individual papers are not always as clear as they could be about which definition they are using, and the different properties of the different definitions are not widely appreciated. This means that different papers may appear at first glance to contradict one another.

In this paper, we aim to clarify the situation by identifying two definitions of retinal vertical disparity that are in common use in the literature. Vertical disparity is sometimes defined as the difference between the elevation-longitude coordinates of the two retinal images of an object, sometimes as the difference in elevation-latitude. Both definitions are valid and sensible, but they have rather different properties, as summarized in Table 1. The differences between the two types of vertical disparity are most significant for objects not at the fixation distance (Figure 7CD), and in the visual periphery. The periphery is where retinal vertical disparities tend to be largest during natural viewing, which has motivated physiologists to investigate vertical disparity tuning there (Durand, et al., 2002 1104). Psychophysically, it has been shown that the perceived depth of centrally viewed disparities (Rogers & Bradshaw, 1993) can be influenced by manipulations of “vertical” disparities in the periphery (i.e. when the field of view is large). Thus, it is particularly important to clarify the difference between the alternative definitions of vertical disparity where stimuli fall on peripheral retina.

For objects close to the fixation point, the images fall close to the fovea in both eyes. Here, latitude and longitude definitions of vertical disparity reduce to the same quantity. In this regime, vertical disparity is much less strongly affected than horizontal disparity by small variations in depth relative to the fixation point. Where this variation is small, it can be treated as independent of surface structure. We have derived expressions giving estimates of each eye position parameter, except elevation, in terms of vertical disparity and its derivatives at the fovea. Although these are only approximations, they perform fairly well in practice (Figure 11). These expressions are closely related to the vertical size ratios discussed in the literature (Backus & Banks, 1999; Backus, et al., 1999; Banks, et al., 2002; Gillam & Lawergren, 1983; Kaneko & Howard, 1996, 1997b; Koenderink & van Doorn, 1976; L. Liu, Stevenson, & Schor, 1994; Rogers & Bradshaw, 1993).

Little if anything in this paper will be new to experts in vertical disparity. However, for the non-cognoscenti, we hope that it may clarify some points that can be confusing. Even for experts, it may serve as a useful reference. We identify, in particular, four areas where we hope this paper makes a useful contribution.

  1. Previous derivations have often been couched in terms of head-centric disparity, or have assumed that the surfaces viewed have special properties such as being oriented vertically. Our derivations are couched entirely in terms of retinal images and does not assume the viewed surface has a particular orientation. We feel this may provide a more helpful mathematical language for describing the properties of disparity encoding in early visual cortex.

  2. We present analytical expressions for both elevation-longitude and elevation-latitude vertical disparity that are valid across the entire retina, for arbitrary gaze angles and cyclotorsion, and for non-zero vertical vergence and cyclovergence. Much previous analysis has relied on parafoveal approximations and has assumed zero vertical vergence, cyclotorsion and cyclovergence.

  3. We present analytical expressions for the average vertical disparity expected at each position in the visual field, up to a scale factor representing the mean convergence.

  4. Explanations relating the perceptual effects of vertical disparity to disparity gradients have sometimes been contrasted with those based on explicit estimates of eye position (Garding, et al., 1995; Longuet-Higgins, 1982; Mayhew & Longuet-Higgins, 1982). This paper is the first to give explicit (though approximate) expressions for 5 binocular eye position parameters in terms of retinal vertical disparity at the fovea. The way in which all 5 eye position parameters can be derived immediately from vertical disparity derivatives has not, as far as we are aware, been laid out explicitly before. Thus, this paper clarifies the underlying unity of gaze-angle and vertical-size-ratio explanations of vertical-disparity illusions such as the induced effect.

Binocular eye position is specified by 6 parameters, 5 of which we have been able to derive from the vertical disparity field around the fovea. The exception is elevation. All the other parameters have a meaning as soon as the two optic centres are defined (and a zero torsion line on the retina), whereas elevation needs an additional external reference frame to say where ‘zero elevation’ is. Disparity is, to first order, independent of how this reference is chosen, meaning that elevation cannot be directly derived from disparity. However, in practice, the visual system obeys Donder's law, meaning that there is a unique relationship between elevation and torsion. The eyes cycloverge inwards when they look down and outwards when they look up (Tweed, 1997c). The torsional states of both eyes can be deduced from the vertical disparity field, as laid out in Equation 24 and Equation 27. Thus, in practice the brain could derive torsion from the gradient of vertical disparity, and use this to obtain an estimate of elevation independent of oculomotor information regarding current eye position (although clearly it would rely on an association between torsion and elevation that would ultimately stem from the oculomotor system). It has already been suggested that the Listing's law relationship between torsion and elevation helps in solving the stereo correspondence problem (Schreiber, et al., 2001; Tweed, 1997c). The fact that it enables elevation to be deduced from the two-dimensional disparity field may be another beneficial side-effect.

Another positive consequence of the framework we have laid out is a set of predictions about the physiological range of retinal disparities. The existing physiological literature does not test such predictions. For example, Durand et al. (2002) explain that “VD is naturally weak in the central part of the visual field and increases with retinal eccentricity”, but then report their results in terms of head-centric Helmholtz disparity (Figure 1A), in which naturally-occurring vertical disparities are always zero, everywhere in the visual field. This makes it impossible to assess whether the results of Durand et al. are consistent with the natural distribution of retinal vertical disparities to which they drew attention in their introduction. This paper has emphasized the importance of calculating neuronal tuning as a function of retinal vertical disparity (whether elevation-longitude or –latitude). Our expressions for average vertical disparity as a function of position in the visual field predict the expected sign of vertical disparity preference. It is intuitively clear that in natural viewing early cortical neurons viewing the top-right visual field should receive a diet of inputs in which the left half-image is higher on the retina than the right (Figure 3), and vice versa for those viewing the top-left visual field. One would expect the tuning of neurons to reflect this biased input. This simple qualitative prediction has not yet been discussed or examined in the physiological literature. Our analysis also makes quantitative predictions. For example, consider eccentricities 5° and 15° in the direction “northeast” from the fovea. Our calculations (Equation 9) predict that the mean vertical disparity tuning of V1 neurons at the larger eccentricity should be 9 times that at the smaller. This too could be tested by appropriate physiological investigations.

There are, of course, limitations to the quantitative predictions we can make from geometric considerations alone. As we have pointed out, to predict the magnitude (and distribution) of expected vertical disparities at any retinal location requires a knowledge of the statistics of binocular eye movements (especially version and vergence) under natural viewing conditions. As recent studies have pointed out, such statistics are quite difficult to gather (Hibbard, 2007; Y. Liu et al., 2008) but they are crucial if the diet of 2D disparities received by binocular neurons across the retina is to be estimated accurately.

Conclusion

The term “vertical disparity” is common in the stereo literature, and the impression is often given that it has an established definition and familiar properties. In fact, neither of these assumptions hold. If the terms “vertical” and “horizontal” are to continue to be used in discussions of binocular disparity, and we argue here that there are reasons in favor of doing so, it is critical that the respective definitions and properties should be set out explicitly, as we have done here.

Supplementary Material

supplementary_mfiles

Figure 10.

Figure 10

Partial differentiation on the retina. The cyclopean retina is shown colored to indicate the value of the vertical disparity field at each point. Differentiating with respect to elevation κ while holding azimuth constant means finding the rate at which vertical disparity changes as one moves up along a line of azimuth-longitude, as shown by the arrow labeled ∂/∂κ. Differentiating with respect to azimuth a, while holding elevation constant, means finding the rate of change as one moves around a line of elevation-latitude.

Acknowledgments

This research was supported by the Royal Society (University Research Fellowship UF041260 to JCAR), MRC (New Investigator Award 80154 to JCAR), the Wellcome Trust (grant 086526/A/08/Z to AG) and the EPSRC (Neuroinformatics Doctoral Training Centre studentship to GPP).

Appendix A: Definitions

Subscripts

Table 2

Symbols

Table 3

Coordinate systems

Head-centered coordinate system (X,Y,Z) for object position in space

Figure 12 shows the head-centered coordinate system used throughout this paper. The X axis points right, the Y axis upwards and the Z axis straight ahead of the observer. By definition, the nodal point of the left eye is at (X,Y,Z) = (i,0,0) and the nodal point of the right eye is at (X,Y,Z) = (−i,0,0), where i represents half the interocular distance I. The position of a point in space can be described as a vector, P=(X,Y,Z).

Figure 12.

Figure 12

Head-centered coordinate system used throughout this paper. The origin is the point midway between the two eyes. The X axis is defined by the nodal points of the two eyes, and points rightwards. The orientation of the XZ plane is defined by primary position, but is approximately horizontal. The Y axis points upwards and the Z axis points in front of the observer.

Eye posture

Each eye has potentially three degrees of freedom, two to specify the gaze direction (azimuth left/right and elevation up/down), and a third to specify the rotation of the eyeball around this axis (torsion). We adopt the Helmholtz coordinate system for describing eye posture (Figure 4). We start with the eye in primary position, looking straight forward so that its optic axis is parallel to the Z axis (Figure 12). We define the torsion here to be zero. To move from this reference state in which all three coordinates are zero to a general posture with torsion, azimuth H and elevation V, we start by rotating the eyeball about the optic axis by the torsion angle T. Next rotate the eye about a vertical axis, i.e. parallel to the Y axis, through the gaze azimuth H. Finally rotate the eye about a horizontal, i.e. interocular axis, through the gaze elevation V. We define these rotation angles to be anti-clockwise around the head-centered coordinate axes. This means that we define positive torsion to be clockwise when viewed from behind the head, positive gaze azimuth to be to the observer's left, and positive elevation to be downwards.

We use the subscripts L and R to indicate the left and right eye (Table 2). Thus, VL is the Helmholtz elevation of the left eye, and VR that of the right eye.

One advantage of Helmholtz coordinates is that it is particularly simple to see whether the eyes are correctly fixating, such that their optic axes intersect at a common fixation point. This occurs if, and only if, the Helmholtz elevations of the two eyes are identical and the optic axes are not diverging. Thus, any difference between VL and VR means that the eyes are misaligned. We refer to this as the vergence error, VR−VL. The difference in the Helmholtz gaze azimuths is the horizontal vergence angle, HR−HL. Negative values mean that the eyes are diverging.

In the mathematical expressions we shall derive below, the vergence angles will usually occur divided by two. We therefore introduce symbols for half the vergence angles. As shown in Table 2, these are indicated with the subscript δ:

Hδ(HRHL)2,etc.

We also introduce cyclopean gaze angles, which are the means of the left and right eyes. As shown in Table 2, these are indicated with the subscript c:

Hc(HR+HL)2.

Rotation matrices

Eye posture can be summarized by a rotation matrix M. So for example if we have a vector which is fixed with respect to the eye, then if the vector is initially r in head-centered coordinates when the eye is in its reference position, it will move to Mr when the eye adopts the posture specified by rotation matrix M. An eye's rotation matrix M depends on the eye's elevation V, gaze azimuth H and torsion T. As above, we use subscripts L and R to indicate the left and right eyes. For the left eye, the rotation matrix is ML = MVLMHLMTL, where

MVL=[1000cosVLsinVL0sinVLcosVL];MHL=[cosHL0sinHL010sinHL0cosHL];MTL=[cosTLsinTL0sinTLcosTL0001] Equation 32

where VL, HL and TL are the gaze elevation, gaze azimuth and torsion of the left eye. The ordering of the matrix multiplication, ML=MVLMHLMTL, is critical, reflecting the definition of the Helmholtz eye coordinates. Obviously, analogous expressions hold for the right eye. Once again, it will be convenient to introduce the cyclopean rotation matrix, which is defined as the mean of the left- and right-eye rotation matrices:

Mc=(MR+ML)2,

and the half-difference rotation matrix:

Mδ=(MRML)2.

It will also be convenient to introduce vectors m which are the columns of these matrices:

mc1=[Mc11Mc21Mc31];mc2=[Mc12Mc22Mc33];mc3=[Mc13Mc23Mc33]
mδ1=[Mδ11Mδ21Mδ31];mc2=[Mδ12Mδ22Mδ33];mc3=[Mδ13Mδ23Mδ33] Equation 33

where Mkl indicates the entry in the kth row and lth column of matrix M.

Gaze-centered coordinate system for object position in space

We use the vectors mck to define a new coordinate system for describing an object's position in space. As well as the head-centered coordinate system (X,Y,Z), we introduce a coordinate system (U,W,S) centered on the direction of cyclopean gaze, as specified by the three Helmholtz angles Hc, Vc and Tc. Whereas Z is the object's distance from the observer measured parallel to the “straight ahead” direction, S is the object's distance parallel to the line of gaze (Figure 5). The coordinates (U,W,S) are defined by writing the vector P=(X,Y,Z) as a sum of the three mc vectors:

P=Umc1+Wmc2+Smc3. Equation 34

Retinal coordinate systems for image position on retina

The retina is at least roughly hemispherical, and treating it as perfectly hemispherical involves no loss of generality, since there is a one-to-one map between a hemisphere and a physiological retina. All the coordinate systems we shall consider are based on the vertical and horizontal retinal meridians. These are great circles on a spherical retina. They are named after their orientations when the eye is in its reference position, looking straight ahead parallel to the Z axis in Figure 12. By definition, our retinal coordinate systems are fixed with respect to the retina, not the head, so as the eye rotates in the head, the “horizontal” and “vertical” meridians will in general no longer be horizontal or vertical with respect to gravity. For this reason we shall call the angle used to specify “horizontal” location the azimuth α, and the angle used to specify “vertical” location, the elevation η. Both azimuth and elevation can be defined as either latitude or longitude. This gives a total of 4 possible retinal coordinate systems (Figure 2B-E). The azimuth-latitude/elevation-longitude coordinate system is the same Helmholtz system we have used to describe eye position. The azimuth-longitude/elevation-latitude coordinate system is the Fick system. One can also choose to use latitude or longitude for both directions. Such azimuth-longitude/elevation-longitude or azimuth-latitude/elevation-latitude systems have the disadvantage that the coordinates become ill-defined around the great circle at 90° to the fovea. However, this is irrelevant to stereopsis, since it is beyond the boundaries of vision. The azimuth-longitude/elevation-longitude coordinate system is very simply related to the Cartesian coordinate system which is standard in the computer vision literature (Figure 2A). We can imagine this as a virtual plane, perpendicular to the optic axis and at unit distance behind the nodal point. To find the image of a point P, we imagine drawing a ray from the point P through the nodal point N and see where this intersects the virtual plane (see Figure 3 of Read & Cumming (2006)). The ray has vector equation p=N+s(P-N), where s represents position along the ray. Points on the retina are given by the vector p= NM + xM + yMŶ , where x and y are the Cartesian coordinates on the planar retina, and the rotation matrix M describes how this plane is rotated with respect to the head. Equating these two expressions for p, we find that

s(PN)=MZ^+xMX^+yMY^

Multiplying the matrix M by the unit vectors simply picks off a column of the matrix, e.g. M = m1. Using this plus the fact that M, MŶ and M are orthonormal, we find that the ray intersects the retina at the Cartesian coordinates

x=m1.(PN)m3.(PN);y=m2.(PN)m3.(PN) Equation 35

It is sometimes imagined that the use of planar retinas involves a loss of generality or is only valid near the fovea, but in fact, no loss of generality is involved, since there is a one-to-one map from the virtual planar retina to the hemispherical retina.

Each coordinate system has a natural definition of “horizontal” or “vertical” disparity associated with it. Disparity is defined to be the difference between the horizontal and vertical coordinates of the two retinal images. So we immediately have three different definitions of retinal vertical disparity: (1) Cartesian vertical disparity, yΔ = yR−yL; (2) elevation-longitude disparity, ηΔ = ηR−ηL; (3) elevation-latitude disparity, κΔ = κR−κL. In Appendix B:, we shall derive expressions for all 3 definitions.

It may also be useful to collect together here for reference the relationships between gaze-centered coordinates and the corresponding retinal coordinates. The equations in Table 4 show where an object located at (U,W,S) in gaze-centered coordinates projects to on the cyclopean retina, in different retinal coordinate systems. Table 5 gives the relationships between location on the retina in different coordinate systems.

Table 4.

The relationship between the quantities (U,W,S), giving an object's location in gaze-centered coordinates (cf Figure 5), and that object's projection onto the cyclopean retina. The projection is given in planar Cartesian coordinates (xc,yc), and as azimuth-longitude αc, elevation-longitude ηc, azimuth-latitude βc and elevation-latitude κc. The object's head-centered coordinates (X,Y,Z) will depend on eye position

U ≈ −Sxc, W ≈ −Syc R2=U2+W2+S2= X2+Y2+Z2
U = −S tan αc W = −S tan κc secαc S = Rcosαccosκc
U = −S tan βc secηc W = −S tan ηc S = Rcosβccosηc.
Table 5.

Relationships between the different retinal coordinate systems shown in Figure 2

Cartesian
(x,y)
(Figure 2A)
Azimuth-longitude, elevation
-longitude
(α, η)
(Figure 2B)
Azimuth-longitude, elevation-
latitude:
(α, κ)
(Fick; Figure 2C)
Azimuth-latitude, elevation-
longitude (β, η)
(Helmholtz; Figure 2D)
Azimuth-latitude, elevation-
latitude: (β, κ) (Figure 2E)
(x,y) x = tan α
y = tan η
x = tan α
y = tan κsecα
x = tan βsecη
y = tan η
x = sinβ / √(cos2κ − sin2β
y = sinκ / √)cos2β − sin2κ
(α, η) α = arctan(x)
η = arctan y
α = α
η=arctan(tanκcosα)
α = arctan(tanβ.secη)
η = η
α = arcsin(sinβsecκ)
η = arcsin(sinκ secβ)
(α, κ) α = arctan(x)
κ=arctan(yx2+1)
α = α
κ = arctan(tan ηcos α)
α=arcsin(sinβcosκ)
κ = κ
α=arcsin(sinβcosκ)
κ = κ
(β, η) β=arctan(xy2+1)
η = arctan(y)
β = arctan(tan αcos η)
η = η
β = arcsin(sin αcos κ)
η = arctan(tanκsecα)
β = β
η=arcsin(sinκcosβ)
(β, κ) β = arctan(x/√(1+y2))
κ = arctan(y/√(1+x2))
β= arctan( tan(α)cos(η))
κ = arctan(tan(η)cos(α))
β = arcsin(cos κsin α)
κ = κ
β = β
κ = arcsin(cos βsin η)

Table 4

Table 5

Appendix B: Derivations

Relationships between the rotation vectors

The fact that rotation matrices are orthogonal means that certain simple relationships hold between the vectors mck and mδk defined in Equation 33 . First, the inner product of any difference vector mδk with the corresponding cyclopean vector mck is identically zero:

mδk.mck=0fork=1,2,3. Equation 36

This is actually a special case of the following more general statement:

mδk.mcl=mδl.mckfork,l=1,2,3. Equation 37

Equation 36 and Equation 37 are exact, and do not depend on any approximations at all.

To obtain the values of these dot products, we need to use Equation 32 to derive expressions for Mc and Mδ in terms of the 6 Helmholtz gaze parameters for the two eyes: HL, VL, TL, HR, VR, TR. We can then use trigonometric identities to re-express these in terms of the cyclopean (half-sum) and vergence (half-difference) equivalents: Hc, Vc, Tc, Hδ, Vδ, Tδ. Needless to say, this yields extremely complicated expressions. However, we now introduce the first critical approximation of this paper. We assume that differences in eye posture are small. We therefore work to first order in the horizontal vergence Hδ, the vertical vergence half-error Vδ, the half-cyclovergence Tδ, i.e. we replace terms like cosHδ with 1, and we neglect terms in sin2Hδ, sinHδ.sinVδ etc. Under these approximations, the 3 mc and the 3 mδ are approximately orthonormal, i.e.

mck.mcl1ifk=land0otherwise;mδk.mδl1ifk=land0otherwise; Equation 38

and we obtain the following simple expressions for inner products of an mc and an m vector:

mδ1.mc2=mδ2.mc1Tδ+VδsinHcmδ2.mc3=mδ3.mc2HδsinTc+VδcosHccosTcmδ1.mc3=mδ3.mc1HδcosTc+VδcosHcsinTc Equation 39

Notice that if the eyes are correctly fixating (Vδ=0) and there is no torsion (Tc=Tδ=0), then the only non-zero inner product is mδ1. mc3 ≈ −Hδ.

Below, we shall also encounter the inner products mc1.X̂,mc2. X̂ and mc3. X̂, where is a unit vector along the X axis. These are the entries in the top row of the cyclopean rotation matrix, which under the above approximation are

mc1.X^=Mc11cosHccosTc;mc2.X^=Mc12sinTccosHc;mc3.X^=Mc13sinHc. Equation 40

We shall also use the following:

2mδ1.PW(TΔ+HcsinVΔ)+S(TccosHΔ+HccosTcsinV)2mδ2.PU(TΔ+HcsinVΔ)+S(TcsinHΔ+HccosTccosVΔ)2mδ3.PU(TccosHΔHccosTcsinVΔ)W(TcsinHΔ+HccosTccosVΔ)

Deriving expressions for retinal disparity

Disparity in Cartesian coordinates on a planar retina

The utility of the above expressions will now become clear. Suppose that an object's position in space is represented by the vector P = (X,Y,Z) in head-centered coordinates. Then the object projects onto the left retina at a point given by (xL,yL) in Cartesian-coordinates, where (Equation 35)

xL=mL1.(PNL)mL3.(PNL);yL=mL2.(PNL)mL3.(PNL)

where NL is the vector from the origin to the nodal point of the left eye, and mLk is the kth column of the left eye's rotation matrix ML. For the left eye, we have NL=i , where is a unit vector along the X axis and i is half the interocular distance, while for the right eye, NR=−i . We shall also rewrite the left and right eye's rotation vectors, mL and mR, in terms of the half-sum and half-difference between the two eyes:

mL=mcmδ;mR=mc+mδ.

The image in the left eye is then

xL=(mc1mδ1).(PiX^)(mc3mδ3).(PiX^), Equation 41

while the expression for the right eye is the same but with the signs of i and mδ reversed:

xR=(mc1mδ1).(PiX^)(mc3mδ3).(PiX^).

Thus, there are two distinct sources of retinal disparity. One of them, the terms in i, arises from the fact that the eyes are in different locations in the head. The other, the terms in mδ, arises from the fact that the eyes may point in different direction. We shall see these two terms emerging in all our future expressions for binocular disparity.

We now make the approximation that both these terms are small, and carry out a Taylor expansion in which we retain only first-order terms of these quantities. To do this, it is helpful to introduce dummy quantities s and j, where mδjsj and i=εj, and the variable ε is assumed to be so small that we can ignore terms in ε2.

xL=(mc1εs1).(PεjX^)(mc3εs3).(PεjX^)mc1.Pmc3.P(1εjmc1.X^mc1.Pεs1.Pmc1.P+εjmc3.X^mc3.P+εs3.Pmc3.P+O(ε2))

Now removing the dummy variables, we have an expression for xL under the small-eye-difference approximation:

xLmc1.Pmc3.P(1+(imc3.X^+mδ3.P)mc3.P(imc1.X^+mδ1.P)mc1.P+O(ε2)). Equation 42

Again, the expression for xR is the same but with the signs of i and rδ reversed. The expressions for y are the same except with subscripts 1 replaced with 2. We can therefore derive the following expressions for the cyclopean position of the image:

xc=xR+xL2mc1.Pmc3.P;yc=yR+yL2mc2.Pmc3.P

while for the Cartesian disparity, we obtain

xΔ=xRxL2mc1.Pmc3.P((imc3.X^+mδ3.P)mc3.P(imc1.X^+mδ1.P)mc1.P)yΔ=yRyL2mc2.Pmc3.P((imc3.X^+mδ3.P)mc3.P(imc2.X^+mδ2.P)mc2.P).

Expressions for mcj.X were given in Equation 40.

Now, instead of specifying P=(X,Y,Z) in head-centered coordinates, I move to the gaze-centered coordinate system (U,W,S) in which an object's position is specified relative to the cyclopean gaze direction (Equation 34):

P=Umc1+Wmc2+Smc3.

Now recall that the inner product of any difference vector mδj with the corresponding cyclopean vector mcj is identically zero (Equation 36). Thus, the term mδ3.P is independent of the object's distance measured along the cyclopean gaze direction, S:

mδ3.P=Umδ3.mc1+Wmδ3.mc2.

Using the relationships between the various m-vectors, Equation 36-Equation 38, we obtain

xcUS,ycWS,

which is in fact obvious given the definition of the cyclopean retina and the cyclopean gaze-centered coordinate system. For the disparity, we obtain

xΔIS2(UMc13SMc11)2S2((U2+S2)mδ1.mc3+UWmδ2.mc3+SWmδ1.mc2)yΔIS2(WMc13SMc12)2S2((W2+S2)mδ2.mc3+UWmδ1.mc3SUmδ1.mc2)

Expressions for the vector inner products, valid under the approximation we are considering, were given in Equation 39 and Equation 40. Substituting these, using the small angle approximation for the δ-quantities, we obtain the following expressions for an object's horizontal and vertical disparity in Cartesian planar coordinates, expressed as a function of its spatial location in gaze-centered coordinates:

xΔIS(USHcsinHccosTccos)+[(U2S2+1)TccosUWS2Tcsin]HΔ[(U2S2+1)HccosTcsin+UWS2HccosTccosWSHcsin]VΔWSTΔyΔIS(WSHcsin+HcsinHccos)+[UWS2Tccos(W2S2+1)Tcsin]HΔ+[USHcsin(W2S2+1)HccosTccosUWS2HccosTccos]VΔ+USTΔ

where to save space we have defined Tcosc=cosTc, etc.

Here, the disparity is expressed as a function of the object's position in space, (U,W,S). But this is not very useful, since the brain has no direct access to this. It is more useful to express disparities in terms of (xc,yc), the position on the cyclopean retina or equivalently the visual direction currently under consideration, together with the distance to the object along the cyclopean gaze, S. Then we obtain the following expressions for an object's horizontal and vertical disparity in Cartesian planar coordinates, expressed as a function of its retinal location in Cartesian planar coordinates:

xΔ(xcHcsin+HccosTccos)IS+[(xc2+1)TccosxcycTcsin]HΔ+[ycHcsin(xc2+1)HccosTcsinxcycHccosTccos]VΔ+ycTΔyΔ(ycHcsinTcsinHccos)IS+[xcycTccos(yc2+1)Tcsin]HΔ[xcHcsin+(yc2+1)HccosTccos+xcycHccosTcsin]VΔxcTΔ

Disparity in retinal longitude

Azimuth-longitude and elevation-longitude on the retina, α and η, are simply related to the planar coordinates x and y:

α=arctan(x);η=arctan(y).

From the approximation to xL given in Equation 42, we have

αLarctan[mc1.Pmc3.P(1+(imc3.X^+mδ3.P)mc3.P(imc1.X^+mδ1.P)mc1.P)].

With the Taylor expansion for arctan, this becomes

αLarctan[mc1.Pmc3.P][(imc3.X^+mδ3.P)mc3.P(imc1.X^+mδ1.P)mc1.P](mc1.P)(mc3.P)(mc3.P)2(mc1.P)2

As before, the analogous expression for αR is the same but with the signs of i and mδ swapped. Thus, we obtain

αcarctan[mc1.Pmc3.P]arctanUS

and

αδ[(imc3.X^+mδ3.P)mc3.P(imc1.X^+mδ1.P)mc1.P](mc1.P)(mc3.P)(mc3.P)2(mc1.P)2.

We similarly obtain for the elevation-longitude cyclopean position and disparity:

ηcarctan[mc2.Pmc3.P]arctanWSηδ[(imc3.X^+mδ3.P)mc3.P(imc2.X^+mδ2.P)mc2.P](mc2.P)(mc3.P)(mc3.P)2(mc2.P)2.

Again substituting for the m, we obtain, in terms of an object's spatial location in gaze-centered coordinates:

αΔ1S2+U2{[UHcsinSHccosTccos]I+[(S2+U2)TccosUWTcsin]HΔ[(S2+U2)HccosTcsin+UWHccosTccos+WSHcsin]VΔWSTΔ}ηΔ1W2+S2{[WHcsin+STcsinHccos]I+[(W2+S2)Tcsin+UWTccos]HΔ[(W2+S2)HccosTccos+UWHccosTcsinUSHcsin]VΔUSTΔ}

We now re-express these disparities in terms of the object's retinal location in azimuth-longitude/elevation-longitude coordinates. From U ≈ −S tanαc, we have

S2S2+U2S2S2+S2tan2αc=cos2αc.

Similarly, S2S2+W2cos2ηc. We therefore arrive at the following expressions for longitude disparity expressed as a function of position on the cyclopean retina in azimuth-longitude/elevation-longitude coordinates:

αΔIScosαc(HccosTccoscosαc+Hcsinsinαc)+[TccossinαccosαctanηcTcsin]HΔ[HccosTcsin+sinαccosαctanηcHccosTccoscos2αctanηcHcsin]VΔ+cos2αctanηcTΔηΔcos2ηc[TcsinHccosHcsintanηc]IS+[tanαcsinηccosηcTccosTcsin]HΔ[HccosTccos+tanαcsinηccosηcHccosTccos+tanαccos2ηcHcsin]VΔtanαccos2ηcTΔ Equation 43

Alternatively, we may wish to express azimuth-longitude disparity as a function of retinal location in an azimuth-longitude/elevation-latitude coordinate system. Elevation-latitude κ is related to azimuth-longitude α and elevation-longitude η as

tanκ=tanη.cosα.

Thus, it is easy to replace ηc in Equation 43 with κc:

αΔIScosαc(HccosTccoscosαc+Hcsinsinαc)+[TccossinαctanκcTcsin]HΔ[HccosTcsin+sinαctanκcHccosTccoscosαctanκcHcsin]VΔ+[cosαctanκc].TΔ

Similarly, we can express elevation-longitude disparity ηΔ as a function of retinal location in an azimuth-latitude/elevation-longitude coordinate system (β,η). Using tanβ = tanαcosη, Equation 43 becomes

ηΔcos2ηc[TcsinHccosHcsintanηc]IS+[tanβcsinηcTccosTcsin]HΔ[HccosTccos+tanβcsinηcHccosTcsin+tanβccosηcHcsin]VΔ+tanβccosηcTΔ

Disparity in retinal latitude

Azimuth- latitude and elevation- latitude on the retina, β and κ, are related to the planar coordinates x and y as

β=arctan(x(y2+1));κ=arctan(y(x2+1)).

From the approximation to xL and yL given in Equation 42, we have

βLarctan[mL1.(PNL)[mL2.(PNL)]2+[mL3.(PNL)]2].

Again doing the Taylor expansion, we obtain the following expressions for horizontal and vertical retinal latitude disparities in terms of an object's spatial location in gaze-centered coordinates:

βΔ=2(Smδ3.mc1+Wmδ2.mc1)W2+S2+I(Mc12UW+Mc13USMc11(W2+S2))(U2+W2+S2)W2+S2κΔ=2(Smδ3.mc2+Wmδ1.mc2)U2+S2+I(Mc11UW+Mc13WSMc12(U2+S2))(U2+W2+S2)U2+S2

Substituting for the various entries in the rotation matrix, Mc, mδ and mc, we obtain

2mδ1.mc2TΔ+HcsinVΔ2mδ2.mc3TcsinHΔ+HccosTccosVΔ2mδ1.mc3TccosHΔ+HccosTcsinVΔβΔ=I(TcsinHccosUWHcsinUS+TccosHccos(W2+S2))(U2+W2+S2)W2+S2(STccosHΔ+SHccosTcsinVΔ+WHcsinVΔ+WTΔ)W2+S2κΔ=I(HccosTccosUW+HcsinWS+TcsinHccos(U2+S2))(U2+W2+S2)U2+S2(STcsinHΔ+SHccosTccosVΔ+UHcsinVΔUTΔ)U2+S2

To express azimuth-latitude disparity as a function of position on the cyclopean retina in azimuth-latitude/elevation-longitude coordinates (βcc), we use the relationships

W=StanηcU=Stanβcsecηc

This yields

βΔ=IScosβccosηc(TcsinHccossinβcsinηc+Hcsinsinβccosηc+TccosHccoscosβc)+(Tccoscosηc)HΔ+cosηc(tanηcHcsinHccosTcsin)VΔ+(sinηc)TΔ

Similarly, if we express elevation-latitude disparity as a function of position on the cyclopean retina in azimuth-longitude/elevation-latitude coordinates (αcc), we obtain

κΔ=IScosαccosκc(TccosHccossinαcsinκcHcsincosαcsinκc+TcsinHccoscosκc)TcsinHΔcosαc(HccosTccoscosαc+sinαcHcsin)VΔsinαcTΔ

This expression simplifies slightly if we replace S, the distance component along the cyclopean line of sight, with R, the shortest distance from the origin to the viewed point. R2=U2+W2+S2, and hence

S=Rcosαccosκc=Rcosβccosηc.

Then we have

κΔ=IR(TccosHccossinαcsinκcHcsincosαcsinκc+TcsinHccoscosκc)TcsinHΔcosαc(HccosTccoscosαc+sinαcHcsin)VΔsinαcTΔ

Appendix C: Tables of expressions for horizontal and vertical disparity in different coordinate systems

C.1: Most general

Table 6

Table 6.

Expressions for horizontal disparity in different coordinate systems. These are correct to first-order in interocular distance I/S (I/R) and in the vergence angles HΔ, VΔ, and TΔ. They hold all over the retina and for any cyclopean gaze Hc, elevation Vc or overall cyclotorsion Tc.

Horizontal disparity Most general expressions
In planar Cartesian
retinal coordinates as
a function of spatial
position in gaze-
centered coordinates
xΔIS(USHcsinHccosTccos)+[(U2S2+1)TccosUWS2Tcsin]HΔ[(U2S2+1)HccosTcsin+UWS2HccosTccos+WSHcsin]VΔWSTΔ
In planar Cartesian
retinal coordinates as
a function of retinal
location in planar Cartesian
coordinates
xΔ(xcHcsin+HccosTccos)IS+[(xc2+1)TccosxcycTcsin]HΔ+[ycHcsin(xc2+1)HccosTcsinxcycHccosTccos]VΔ+ycTΔ
In azimuth-longitude,
as a function of spatial
location in gaze-
centered coordinates
αΔ1S2+U2{[UHcsinSHccosTccos]I+[(S2+U2)TccosUWTcsin]HΔ[(S2+U2)HccosTcsin+UWHccosTccos+WSHcsin]VΔWSTΔ}
In azimuth-longitude,
as a function of retinal
location in azimuth-
longitude/elevation-
longitude coordinates
αΔIScosαc(HccosTccoscosαc+Hcsinsinαc)+[TccossinαccosαctanηcTcsin]HΔ[HccosTcsin+sinαccosαctanηcHccosTccoscos2αctanηcHcsin]VΔ+cos2αctanηcTΔ
In azimuth-longitude,
as a function of retinal
location in azimuth-
longitude/elevation-
latitude coordinates
αΔIScosαc(HccosTccoscosαc+Hcsinsinαc)+(TccossinαctanκcTcsin)HΔ(HccosTcsin+TccosHccossinαctanκcHcsincosαctanκc)VΔ+(cosαctanκc)TΔαΔIRsecκc(HccosTccoscosαc+Hcsinsinαc)+(TccossinαctanκcTcsin)HΔ(HccosTcsin+TccosHccossinαctanκcHcsincosαctanκc)VΔ+(cosαctanκc)TΔ
In azimuth-latitude, as
a function of spatial
location in gaze-
centered coordinates
βΔ=I(TcsinHccosUWHcsinUS+TccosHccos(W2+S2))(U2+W2+S2)W2+S2(STccosHΔ+SHccosTcsinVΔ+WHcsinVΔ+WTΔ)W2+S2
In azimuth-latitude, as
a function of retinal
location in azimuth-
latitude/elevation-
longitude coordinates
βΔ=IR(TcsinHccossinβcsinηc+Hcsinsinβccosηc+TccosHccoscosβc)+(Tccoscosηc)HΔ+cosηc(tanηcHcsinHccosTcsin)VΔ+(sinηc)TΔ

Table 7

The expressions in Table 6 and Table 7 assume that the cyclovergence between the eyes, TΔ, is small. They do not assume anything about the overall cyclotorsion, Tc. Cyclotorsion rotates the eyes in the head, mixing up vertical and horizontal disparity. This can occur when the head tilts over, so that the interocular axis is no longer horizontal with respect to gravity. In the extreme case of Tc=90°, the vertical and horizontal directions have actually swapped over (y→x and x→−y). One can verify from the above results that the expressions for vertical and horizontal disparity also swap over (i.e. xΔ with Tc=90° is the same as yΔ with Tc=0, after replacing y with x and x with −y), a quick “sanity check” on the results.

C.2 Zero overall cyclotorsion

Table 8

Table 8.

Expressions for horizontal disparity in different coordinate systems. These are correct to first-order in interocular distance I/S (I/R) and in the vergence angles HΔ, VΔ, and TΔ. They hold all over the retina and for any cyclopean gaze Hc or elevation Vc, provided there is no overall cyclotorsion, Tc=0.

Horizontal disparity With zero overall cyclotorsion, Tc=0
In planar Cartesian retinal
coordinates as a
function of spatial position
in gaze-centered
coordinates
xΔIS(USHcsinHccos)+(U2S2+1)HΔ[UWS2Hccos+WSHcsin]VΔWSTΔ
In planar Cartesian retinal
coordinates as a
function of retinal location
in planar Cartesian
coordinates
xΔ(xcHcsin+Hccos)IS+(xc2+1)HΔ+(ycHcsinxcycHccos)VΔ+ycTΔ
In azimuth-longitude, as
a function of spatial location
in gaze-centered
coordinates
αΔ1S2+U2{[UHcsinSHccos]I+(S2+U2)HΔ[UWHccos+WSHcsin]VΔWSTΔ}
In azimuth-longitude, as
a function of retinal location
in azimuth-
longitude/elevation-
longitude coordinates
αΔIScosαccos(Hcαc)+HΔ+VΔcosαctanηcsin(Hcαc)+TΔcos2αctanηc
In azimuth-longitude, as
a function of retinal location
in azimuth-
longitude/elevation-
latitude coordinates
αΔIScosαccos(Hcαc)+HΔ+VΔtanκcsin(Hcαc)+TΔcosαctanκcαΔIRsecκccos(Hcαc)+HΔ+VΔtanκcsin(Hcαc)+(cosαctanκc)TΔ
In azimuth-latitude, as a
function of spatial location
in gaze-centered
coordinates
βΔ=I(HcsinUS+Hccos(W2+S2))(U2+W2+S2)W2+S2(SHΔ+WHcsinVΔ+WTΔ)W2+S2
In azimuth-latitude, as a
function of retinal location
in azimuth-
latitude/elevation-
longitude coordinates
βΔ=IR(Hcsinsinβccosηc+Hccoscosβc)+HΔTccoscosηc+VΔHcsincosηctanηc+TΔsinηc

Table 9

Table 9.

Expressions for vertical disparity in different coordinate systems. These are correct to first-order in interocular distance I and in the vergence angles HΔ, VΔ, and TΔ. They hold all over the retina and for any cyclopean gaze Hc or elevation Vc, provided there is no overall cyclotorsion, Tc=0.

Vertical disparity With zero overall cyclotorsion, Tc=0
In planar Cartesian retinal coordinates
as a function of spatial position
in gaze-centered coordinates
yΔIWS2Hcsin+HΔUWS2+[USHcsin(W2S2+1)Hccos]VΔ+USTΔ
In planar Cartesian retinal coordinates
as a function of retinal location
in planar Cartesian coordinates
yΔISycHcsin+HΔxcyc[xcHcsin+(yc2+1)Hccos]VΔxcTΔ
In elevation-longitude, as a function
of spatial location in gaze-
centered coordinates
ηΔ1W2+S2{[WHcsin]I+UWHΔ[(W2+S2)HccosUSHcsin]VΔ+USTΔ}
In elevation -longitude, as a function
of retinal location in azimuth-
longitude/elevation-longitude coordinates
ηΔISHcsinsinηccosηc+HΔtanαcsinηccosηcVΔ(Hccos+tanαccos2ηcHcsin)tanαccos2ηcTΔ
In elevation -longitude, as a function
of retinal location in azimuth-
latitude/elevation-longitude coordinates
ηΔISHcsinsinηccosηc+HΔtanβcsinηcVΔ(Hccos+tanβccosηcHcsin)tanβccosηcTΔηΔIRHcsinsinηccosβc+HΔtanβcsinηcVΔ(Hccos+tanβccosηcHcsin)tanβccosηcTΔ
In elevation-latitude, as a function
of spatial location in gaze-centered
coordinates
κΔ=I(HccosUW+HcsinWS)(U2+W2+S2)U2+S2(SHccosVΔUHcsinVΔUTΔ)U2+S2
In elevation-latitude, as a function
of retinal location in azimuth- longitude
/elevation- latitude coordinates
κΔ=IRsinκcsin(Hcαc)VΔcos(Hcαc)sinαcTΔ

C.3 Zero overall cyclotorsion, cyclovergence and vertical vergence error

Table 10

Table 11

Table 11.

Expressions for vertical disparity in different coordinate systems. These are correct to first-order in interocular distance I/S (I/R) and in the convergence angle HΔ. They assume cyclotorsion, cyclovergence and vertical vergence are all zero: Tc=TΔ=VΔ=0. They hold all over the retina and for any cyclopean gaze Hc or elevation Vc.

Vertical disparity For zero torsion and vertical vergence error
In planar Cartesian retinal coordinates as a function of retinal
location in planar Cartesian coordinates:
yΔyc(HcsinIS+xcHΔ)
In planar Cartesian retinal coordinates as a function of spatial
position in gaze-centered coordinates
yΔWS2(IHcsin+UHΔ)
In elevation-longitude, as a function of spatial location in gaze-
centered coordinates
ηΔWW2+S2(IsinHc+UHΔ)
In elevation-longitude, as a function of retinal location in azimuth-
longitude/elevation-longitude coordinates
ηΔsinηccosηc(ISsinHc+HΔtanαc)
In elevation -longitude, as a function of retinal location in azimuth-
latitude/elevation-longitude coordinates
ηΔsinηc(ISsinHccosηc+HΔtanβc)
In elevation-latitude, as a function of spatial location in gaze-
centered coordinates
κΔ=IW(UcosHc+SsinHc)(U2+W2+S2)U2+S2
In elevation-latitude, as a function of retinal location in azimuth-
longitude/elevation- latitude coordinates
κΔ=ISsinκccosκccosαcsin(αcHc)κΔ=IRsinκcsin(αcHc)

Footnotes

Commercial relationships: none.

Contributor Information

Jenny C. A. Read, Institute of Neuroscience, Newcastle University

Graeme P. Phillipson, Neuroinformatics Doctoral Training Centre, Institute of Adaptive and Neural Computation, School of Informatics, Edinburgh University

Andrew Glennerster, School of Psychology and Clinical Language Sciences, Reading University.

References

  1. Adams W, Frisby JP, Buckley D, Garding J, Hippisley-Cox SD, Porrill J. Pooling of vertical disparities by the human visual system. Perception. 1996;25(2):165–176. doi: 10.1068/p250165. [DOI] [PubMed] [Google Scholar]
  2. Allison RS, Rogers BJ, Bradshaw MF. Geometric and induced effects in binocular stereopsis and motion parallax. Vision Res. 2003;43(17):1879–1893. doi: 10.1016/s0042-6989(03)00298-0. [DOI] [PubMed] [Google Scholar]
  3. Backus BT, Banks MS. Estimator reliability and distance scaling in stereoscopic slant perception. Perception. 1999;28(2):217–242. doi: 10.1068/p2753. [DOI] [PubMed] [Google Scholar]
  4. Backus BT, Banks MS, van Ee R, Crowell JA. Horizontal and vertical disparity, eye position, and stereoscopic slant perception. Vision Res. 1999;39(6):1143–1170. doi: 10.1016/s0042-6989(98)00139-4. [DOI] [PubMed] [Google Scholar]
  5. Banks MS, Backus BT. Extra-retinal and perspective cues cause the small range of the induced effect. Vision Res. 1998;38(2):187–194. doi: 10.1016/s0042-6989(97)00179-x. [DOI] [PubMed] [Google Scholar]
  6. Banks MS, Backus BT, Banks RS. Is vertical disparity used to determine azimuth? Vision Res. 2002;42(7):801–807. doi: 10.1016/s0042-6989(01)00283-8. [DOI] [PubMed] [Google Scholar]
  7. Banks MS, Hooge IT, Backus BT. Perceiving slant about a horizontal axis from stereopsis. J Vis. 2001;1(2):55–79. doi: 10.1167/1.2.1. [DOI] [PubMed] [Google Scholar]
  8. Barlow HB, Blakemore C, Pettigrew JD. The neural mechanisms of binocular depth discrimination. J Physiol. 1967;193:327–342. doi: 10.1113/jphysiol.1967.sp008360. [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Berends EM, Erkelens CJ. Strength of depth effects induced by three types of vertical disparity. Vision Res. 2001;41(1):37–45. doi: 10.1016/s0042-6989(00)00215-7. [DOI] [PubMed] [Google Scholar]
  10. Berends EM, van Ee R, Erkelens CJ. Vertical disparity can alter perceived direction. Perception. 2002;31(11):1323–1333. doi: 10.1068/p3440. [DOI] [PubMed] [Google Scholar]
  11. Bishop PO. Vertical disparity, egocentric distance and stereoscopic depth constancy: a new interpretation. Proc R Soc Lond B Biol Sci. 1989;237(1289):445–469. doi: 10.1098/rspb.1989.0059. [DOI] [PubMed] [Google Scholar]
  12. Bishop PO, Kozak W, Vakkur GJ. Some quantitative aspects of the cat's eye: axis and plane of reference, visual field co-ordinates and optics. J Physiol. 1962;163:466–502. doi: 10.1113/jphysiol.1962.sp006990. [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Brenner E, Smeets JB, Landy MS. How vertical disparities assist judgements of distance. Vision Res. 2001;41(25-26):3455–3465. doi: 10.1016/s0042-6989(01)00206-1. [DOI] [PubMed] [Google Scholar]
  14. Carpenter RH. Movements of the eyes. 2nd ed. Pion Ltd; 1988. [Google Scholar]
  15. Clement RA. Gaze angle explanations of the induced effect. Perception. 1992;21(3):355–357. doi: 10.1068/p210355. [DOI] [PubMed] [Google Scholar]
  16. Cumming BG. An unexpected specialization for horizontal disparity in primate primary visual cortex. Nature. 2002;418(6898):633–636. doi: 10.1038/nature00909. [DOI] [PubMed] [Google Scholar]
  17. Cumming BG, Johnston EB, Parker AJ. Vertical disparities and perception of three-dimensional shape. Nature. 1991;349(6308):411–413. doi: 10.1038/349411a0. [DOI] [PubMed] [Google Scholar]
  18. Duke PA, Howard IP. Vertical-disparity gradients are processed independently in different depth planes. Vision Res. 2005;45(15):2025–2035. doi: 10.1016/j.visres.2005.01.027. [DOI] [PubMed] [Google Scholar]
  19. Durand JB, Celebrini S, Trotter Y. Neural Bases of Stereopsis across Visual Field of the Alert Macaque Monkey. Cereb Cortex. 2006 doi: 10.1093/cercor/bhl050. [DOI] [PubMed] [Google Scholar]
  20. Durand JB, Zhu S, Celebrini S, Trotter Y. Neurons in parafoveal areas V1 and V2 encode vertical and horizontal disparities. J Neurophysiol. 2002;88(5):2874–2879. doi: 10.1152/jn.00291.2002. [DOI] [PubMed] [Google Scholar]
  21. Erkelens CJ, van Ee R. A computational model of depth perception based on headcentric disparity. Vision Res. 1998;38(19):2999–3018. doi: 10.1016/s0042-6989(98)00084-4. [DOI] [PubMed] [Google Scholar]
  22. Friedman RB, Kaye MG, Richards W. Effect of vertical disparity upon stereoscopic depth. Vision Res. 1978;18(3):351–352. doi: 10.1016/0042-6989(78)90172-4. [DOI] [PubMed] [Google Scholar]
  23. Frisby JP, Buckley D, Grant H, Garding J, Horsman JM, Hippisley-Cox SD, et al. An orientation anisotropy in the effects of scaling vertical disparities. Vision Res. 1999;39(3):481–492. doi: 10.1016/s0042-6989(98)00126-6. [DOI] [PubMed] [Google Scholar]
  24. Garding J, Porrill J, Mayhew JE, Frisby JP. Stereopsis, vertical disparity and relief transformations. Vision Res. 1995;35(5):703–722. doi: 10.1016/0042-6989(94)00162-f. [DOI] [PubMed] [Google Scholar]
  25. Gillam B, Chambers D, Lawergren B. The role of vertical disparity in the scaling of stereoscopic depth perception: an empirical and theoretical study. Percept Psychophys. 1988;44(5):473–483. doi: 10.3758/bf03210433. [DOI] [PubMed] [Google Scholar]
  26. Gillam B, Lawergren B. The induced effect, vertical disparity, and stereoscopic theory. Percept Psychophys. 1983;34(2):121–130. doi: 10.3758/bf03211336. [DOI] [PubMed] [Google Scholar]
  27. Gonzalez F, Justo MS, Bermudez MA, Perez R. Sensitivity to horizontal and vertical disparity and orientation preference in areas V1 and V2 of the monkey. Neuroreport. 2003;14(6):829–832. doi: 10.1097/00001756-200305060-00010. [DOI] [PubMed] [Google Scholar]
  28. Hartley R, Zisserman A. Multiple view geometry in computer vision. Cambridge University Press; Cambridge, UK: 2000. [Google Scholar]
  29. Helmholtz H. v. Treatise on physiological optics. Optical Society of America; Rochester, NY: 1925. [Google Scholar]
  30. Hibbard PB. A statistical model of binocular disparity. Visual Cognition. 2007;15(2):149–165. [Google Scholar]
  31. Howard IP. Seeing in depth, volume 1: Basic mechanisms. I. Porteous; Ontario, Canada: 2002. [Google Scholar]
  32. Howard IP, Allison RS, Zacher JE. The dynamics of vertical vergence. Exp Brain Res. 1997;116(1):153–159. doi: 10.1007/pl00005735. [DOI] [PubMed] [Google Scholar]
  33. Howard IP, Rogers BJ. Seeing in depth, volume 2: Depth perception. I. Porteous; Ontario, Canada: 2002. [Google Scholar]
  34. Kaneko H, Howard IP. Relative size disparities and the perception of surface slant. Vision Res. 1996;36(13):1919–1930. doi: 10.1016/0042-6989(95)00258-8. [DOI] [PubMed] [Google Scholar]
  35. Kaneko H, Howard IP. Spatial limitation of vertical-size disparity processing. Vision Res. 1997a;37(20):2871–2878. doi: 10.1016/s0042-6989(97)00099-0. [DOI] [PubMed] [Google Scholar]
  36. Kaneko H, Howard IP. Spatial properties of shear disparity processing. Vision Res. 1997b;37(3):315–323. doi: 10.1016/s0042-6989(96)00165-4. [DOI] [PubMed] [Google Scholar]
  37. Koenderink JJ, van Doorn AJ. Geometry of binocular vision and a model for stereopsis. Biol Cybern. 1976;21(1):29–35. doi: 10.1007/BF00326670. [DOI] [PubMed] [Google Scholar]
  38. Liu L, Stevenson SB, Schor CM. A polar coordinate system for describing binocular disparity. Vision Res. 1994;34(9):1205–1222. doi: 10.1016/0042-6989(94)90302-6. [DOI] [PubMed] [Google Scholar]
  39. Liu Y, Bovik AC, Cormack LK. Disparity statistics in natural scenes. J Vis. 2008;8(11):19 11–14. doi: 10.1167/8.11.19. [DOI] [PubMed] [Google Scholar]
  40. Longuet-Higgins HC. A computer algorithm for reconstructing a scene from two projections. Nature. 1981;293:133–135. [Google Scholar]
  41. Longuet-Higgins HC. The role of the vertical dimension in stereoscopic vision. Perception. 1982;11(4):377–386. doi: 10.1068/p110377. [DOI] [PubMed] [Google Scholar]
  42. Matthews N, Meng X, Xu P, Qian N. A physiological theory of depth perception from vertical disparity. Vision Res. 2003;43(1):85–99. doi: 10.1016/s0042-6989(02)00401-7. [DOI] [PubMed] [Google Scholar]
  43. Mayhew JE. The interpretation of stereo-disparity information: the computation of surface orientation and depth. Perception. 1982;11(4):387–403. doi: 10.1068/p110387. [DOI] [PubMed] [Google Scholar]
  44. Mayhew JE, Longuet-Higgins HC. A computational model of binocular depth perception. Nature. 1982;297(5865):376–378. doi: 10.1038/297376a0. [DOI] [PubMed] [Google Scholar]
  45. Minken AW, Van Gisbergen JA. A three-dimensional analysis of vergence movements at various levels of elevation. Exp Brain Res. 1994;101(2):331–345. doi: 10.1007/BF00228754. [DOI] [PubMed] [Google Scholar]
  46. Mok D, Ro A, Cadera W, Crawford JD, Vilis T. Rotation of Listing's plane during vergence. Vision Res. 1992;32(11):2055–2064. doi: 10.1016/0042-6989(92)90067-s. [DOI] [PubMed] [Google Scholar]
  47. Ogle KN. Space perception and vertical disparity. J Opt Soc Am. 1952;42(2):145–146. [PubMed] [Google Scholar]
  48. Pierce BJ, Howard IP. Types of size disparity and the perception of surface slant. Perception. 1997;26(12):1503–1517. doi: 10.1068/p261503. [DOI] [PubMed] [Google Scholar]
  49. Read JCA, Cumming BG. Understanding the cortical specialization for horizontal disparity. Neural Comput. 2004;16:1983–2020. doi: 10.1162/0899766041732440. [DOI] [PMC free article] [PubMed] [Google Scholar]
  50. Read JCA, Cumming BG. Does visual perception require vertical disparity detectors? Journal of Vision. 2006;6(12):1323–1355. doi: 10.1167/6.12.1. [DOI] [PubMed] [Google Scholar]
  51. Rogers BJ, Bradshaw MF. Vertical disparities, differential perspective and binocular stereopsis. Nature. 1993;361(6409):253–255. doi: 10.1038/361253a0. [DOI] [PubMed] [Google Scholar]
  52. Rogers BJ, Bradshaw MF. Disparity scaling and the perception of frontoparallel surfaces. Perception. 1995;24(2):155–179. doi: 10.1068/p240155. [DOI] [PubMed] [Google Scholar]
  53. Rogers BJ, Cagenello R. Disparity curvature and the perception of three-dimensional surfaces. Nature. 1989;339(6220):135–137. doi: 10.1038/339135a0. [DOI] [PubMed] [Google Scholar]
  54. Rogers BJ, Koenderink J. Monocular aniseikonia: a motion parallax analogue of the disparity-induced effect. Nature. 1986;322(6074):62–63. doi: 10.1038/322062a0. [DOI] [PubMed] [Google Scholar]
  55. Schreiber K, Crawford JD, Fetter M, Tweed D. The motor side of depth vision. Nature. 2001;410(6830):819–822. doi: 10.1038/35071081. [DOI] [PubMed] [Google Scholar]
  56. Serrano-Pedraza I, Phillipson GP, Read JCA. A specialization for relative vertical disparity. submitted to Journal of Vision. 2009 doi: 10.1167/10.3.2. [DOI] [PubMed] [Google Scholar]
  57. Serrano-Pedraza I, Read JCA. Stereo vision requires an explicit encoding of vertical disparity. Journal of Vision. 2009;9(4):1–13. doi: 10.1167/9.4.3. [DOI] [PubMed] [Google Scholar]
  58. Somani RA, DeSouza JF, Tweed D, Vilis T. Visual test of Listing's law during vergence. Vision Res. 1998;38(6):911–923. doi: 10.1016/s0042-6989(97)00228-9. [DOI] [PubMed] [Google Scholar]
  59. Stenton SP, Frisby JP, Mayhew JE. Vertical disparity pooling and the induced effect. Nature. 1984;309(5969):622–623. doi: 10.1038/309622a0. [DOI] [PubMed] [Google Scholar]
  60. Stevenson SB, Schor CM. Human stereo matching is not restricted to epipolar lines. Vision Res. 1997;37(19):2717–2723. doi: 10.1016/s0042-6989(97)00097-7. [DOI] [PubMed] [Google Scholar]
  61. Tweed D. Kinematic principles of three-dimensional gaze control. In: Fetter M, Haslwanter TP, Misslisch H, Tweed D, editors. Three-dimensional kinematics of eye, head and limb movements. Harwood Academic Publishers; Amsterdam: 1997a. [Google Scholar]
  62. Tweed D. Three-dimensional model of the human eye-head saccadic system. J Neurophysiol. 1997b;77(2):654–666. doi: 10.1152/jn.1997.77.2.654. [DOI] [PubMed] [Google Scholar]
  63. Tweed D. Visual-motor optimization in binocular control. Vision Res. 1997c;37(14):1939–1951. doi: 10.1016/s0042-6989(97)00002-3. [DOI] [PubMed] [Google Scholar]
  64. Van Rijn LJ, Van den Berg AV. Binocular eye orientation during fixations: Listing's law extended to include eye vergence. Vision Res. 1993;33(5-6):691–708. doi: 10.1016/0042-6989(93)90189-4. [DOI] [PubMed] [Google Scholar]
  65. Westheimer G. Vertical disparity detection: is there an induced size effect? Invest Ophthalmol Vis Sci. 1978;17(6):545–551. [PubMed] [Google Scholar]
  66. Williams TD. Vertical disparity in depth perception. Am J Optom Arch Am Acad Optom. 1970;47(5):339–344. doi: 10.1097/00006324-197005000-00001. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

supplementary_mfiles

RESOURCES