Abstract
Geometry is closely linked to visual perception; yet, very little is known about the geometry of visual processing beyond early retinotopic organization. We present a variety of perceptual phenomena showing that a retinotopic representation is neither sufficient nor necessary to support form perception. We discuss the popular “object files” concept as a candidate for non-retinotopic representations and, based on its shortcomings, suggest future directions for research using local manifold representations. We suggest that these manifolds are created by the emergence of dynamic reference-frames that result from motion segmentation. We also suggest that the metric of these manifolds is based on relative motion vectors.
Index Terms: Non-retinotopic perception, Object files, Reference frames, Retinotopy, Visual system
I. Introduction
There exists a fundamental relationship between visual perception and geometry. Points, lines, angles, surfaces, shapes, and solids constitute elementary objects of both. They are both concerned with the measurement and characterization of these elements as well as their relationships. It seems reasonable therefore that geometry would be the appropriate branch of mathematics in understanding how information is represented and processed in the visual system. The geometry of early visual processing is relatively well understood. Three-dimensional stimuli are imaged on the retina through the optics of the eye following the laws of optics. This geometric transformation maps neighboring points in the environment to neighboring photoreceptors in the retina. The projections of neurons from retina to early visual cortical areas preserve these neighborhood relations, a property known as retinotopy. Thus, scenes create two-dimensional images like images created by a camera.
From the reverse-engineering perspective, computational theories of visual processing rely on these retinotopic neighborhood relationships. For example, in these theories, the processing of stimulus attributes such as orientation, contour, and shape is based on neighborhood relationships between points, lines, etc. Neural counterparts of these theories are based on processing carried out by neurons whose receptive fields reflect retinotopically-based computations that lead to edge detection, contour completion, etc. As an example, Fig. 1 shows schematically how local interactions between retinotopically organized neurons in the primary visual cortex can lead to the detection of oriented edges in the stimulus. Empirical studies also use extensively the knowledge of this retinotopic organization. For example, investigations, carried out at small-scale, probe the activities of a few neurons by single or small arrays of electrodes inserted into the cortex. By moving the stimulus in a small spatial neighborhood, one can probe interactions between the recorded cell and its neighbors and thereby map a “receptive field” for the cell which can be interpreted as local interactions between neurons in the retinotopic space. Large-scale investigations using the functional Magnetic Resonance Imaging (fMRI) technique routinely use retinotopic mapping approaches (Fig. 2) to determine the geometric correspondence between visual space and neural maps as well as to determine borders between different visual areas (e.g., V1, V2, V3, etc.) in the cortex.
From the forward-engineering perspective, the knowledge of retinotopic representations is key in designing sophisticated retinal or cortical implants so as to match the geometry of implant electrodes to that of the nervous system.
While retinotopic representations in early visual areas are well established, very little is known about the organization of higher visual areas; although some recent studies revealed retinotopic maps driven by stimulus or attention beyond the early visual cortex [3], [72].
Regardless how far retinotopic representations persist in the visual cortex, based on psychophysical data, it has been known for more than a century that a retinotopic representation is neither sufficient nor necessary for the perception of spatially extended form. Therefore, a fundamental problem in neuroscience is to understand the geometrical properties of visual processing from the early retinotopic to the higher non-retinotopic levels. In the following, we will review evidence highlighting the shortcomings of retinotopic representations and point to future research directions that may be instrumental in revealing the geometry of visual representations beyond the basic retinotopic maps.
II. A Retinotopic Representation is not Sufficient for form perception
Metacontrast masking refers to the reduced visibility of one stimulus, called the target, due to the presence of a second, spatially non-overlapping, stimulus called the “mask” (Fig. 3) [7],[12]. The presentation of the mask is delayed with respect to the presentation of the target. Although the target is fully visible when presented without the mask, the spatially and thus retinotopically, non-overlapping mask can render it completely invisible. This empirical observation shows that the presence of a retinotopic image is not a sufficient condition for the perception of form. It also highlights the importance of the dynamic context in determining whether form perception takes place. In the following sub-sections, we discuss two dynamic aspects of natural viewing conditions that suggest why a retinotopic representation is insufficient in the computation of visual form.
A. Eye movements
Natural vision is highly dynamic in the sense that we continuously explore our environment by frequent gaze shifts. Our eyes undertake both voluntary and involuntary movements which can be conjunctive (the two eyes move in the same direction) or disjunctive (the two eyes move in opposite direction), smooth (for example, when we pursue a smoothly moving object) or non-smooth (for example when we make a “saccadic eye movement” by shifting our gaze rapidly from one fixation point to another). All these observations indicate that the retinotopic stimulus undergoes drastic and complex changes during normal vision, 3–4 times per second. Nevertheless, our perception of the world remains stable suggesting that the visual system transforms retinotopic representations to those that are invariant with respect to eye movements.
B. Movement of objects
A briefly presented stimulus remains visible after the stimulus offset, a phenomenon known as visible persistence. Under normal viewing conditions, visible persistence is approximately 120ms. Based on this value of visible persistence, one would expect moving objects to appear highly smeared, much like pictures of moving objects taken by a camera with a shutter speed of 1/120ms. Fig. 4 shows an illustrative picture where static objects appear clear while moving objects are highly blurred. Yet, under normal viewing conditions, moving objects appear relatively sharp and clear. This raises two questions: How is motion smear avoided and how is clarity of form achieved?
With respect to the first question, as mentioned above, metacontrast masking operating in the retinotopic space can reduce the spatial extent of motion smear by suppressing activity underlying the motion smear [20],[64]. However, metacontrast mechanisms address only the first question stated above. If we consider Fig. 4, retinotopic metacontrast mechanisms would reduce the spatial extent of motion streaks thereby reducing the amount of blur in the picture. However, moving objects would lack clarity and thus would still suffer from having a “ghost-like” appearance. For example, in Fig. 4, stationary objects appear relatively clear with well-defined form. In contrast, moving objects are fuzzy and lack clarity of form. This is particularly true for objects that move fast in the image plane, or equivalently in retinotopic representations (e.g., the vehicle close to the observer crossing the street). This is because static objects remain long enough on a fixed region of the film to expose sufficiently the chemicals while moving objects expose each part of the film only briefly thus failing to provide sufficient exposure to any specific part of the film. A similar argument can be made for retinotopic representations. A moving object will stimulate each retinotopically localized neuron briefly resulting in an incompletely processed form information that is spread across the retinotopic space just like the ghostly appearances in Fig. 4. We call this fundamental problem of moving form computation “the problem of moving ghosts”. As we will discuss in the following sections, several lines of evidence suggest that the visual system overcomes this “problem of moving ghosts” by computing the form of moving objects in non-retinotopic representations.
III. A Retinotopic Representation is not necessary for form perception
The geometric transformation of the three-dimensional environment into a two-dimensional image implies that objects can partly or fully occlude each other in the retinotopic representations. As a result, retinotopic information about the occluded object will be incomplete; this can be partial or complete, steady (for static objects) or transient (for dynamic objects). In the presence of occlusions, we do not perceive a set of fragmented parts; rather, the occluded object appears as a whole, a phenomenon known as amodal completion [44]. For example, in Fig. 5, the three dark segments appear as parts of the same single object, an occluded triangle. While one can argue that amodal completion can be accomplished in retinotopic representations, a limiting case of dynamic occlusion where a moving object is viewed behind a narrow slit shows that a retinotopic image is not necessary for the perception of spatially extended form. This paradigm, called anorthoscopic perception, derives its name from the anorthoscope, a device invented by Plateau [63]. Since its invention, the anorthoscope has been used in several laboratories as a scientific equipment to study human perception (e.g., [32],[66],[70],[92]). Fig. 6 depicts the retinotopic representations generated in anorthoscopic viewing. Given the optics of the eye, all information about the moving object is mapped in a very narrow retinotopic region corresponding to the interior of the slit. In other words, there is no spatially extended retinotopic representation for the moving stimulus. Yet, observers perceive a spatially extended shape moving behind the slit rather than an incoherent pattern confined to the region of the slit. Can this be explained based on retinotopic representations? Helmholtz [32] put forward the “retinal painting” hypothesis to explain anorthoscopic perception. According to this hypothesis, if the eyes move smoothly during the presentation of the stimulus, then successive parts of the stimulus shown behind the slit will be imaged on adjacent retinotopic positions. Hence, a retinotopic picture of the figure will be “painted” as a consequence of eye movements. This explanation implies that retinotopic representations are sufficient to explain anorthoscopic perception. While it is true that retinal painting can give rise to the perception of form, several studies showed that anorthoscopic perception does occur without any contribution of eye movements (e.g. [27],[46]). Taken together, these results show that a retinotopic image is not necessary for the perception of spatially extended form and suggest that the visual system must be using non-retinotopic representations for spatially extended form. Thus understanding mechanisms of anorthoscopic perception can play a crucial role not only in revealing the characteristics of non-retinotopic representations used by the visual system but also the transformations that occur between early retinotopic representations and higher-level non-retinotopic representations.
Before we analyze anorthoscopic perception further, let us consider the geometrical aspects of an extensively studied function of the visual system, viz., the synthesis of invariant “object” representations and discrimination of different objects or patterns
IV. Insights on non-retinotopic representations from invariant pattern recognition and discrimination
From the mathematical point of view, natural visual stimuli reside in a very high-dimensional space and the probability of receiving the exact same stimulus more than once is quite low. For example, the proximal stimulus corresponding to a face can change over time due to changes in position, size, pose, expression, lighting, and occlusions to name just a few factors. Notwithstanding these variations, the visual system is capable of building more abstract representations that allow the integration of different instances of an “object” into a single category and, at the same time, segregation of instances of different objects into different categories. These dual problems of pattern recognition and discrimination attracted extensive interest from both biological and computer vision communities. Rosenblatt’s “Perceptron” is one of the earliest examples of biologically-inspired and mathematically-formulated pattern-recognition systems [70]. Interestingly, a geometrical analysis of elementary versions of Perceptron revealed that their scope is limited to linearly-separable problems [45]. The geometrical analysis also reveals that a cascade of such networks can solve arbitrarily complex pattern discrimination problems. They achieve this feat by transforming the geometrical representation of the inputs into linearly-separable internal representations [24]. Another key aspect of the geometrical transformations applied to inputs is the ability to reduce the dimensionality of the stimuli. Several methods for embedding high-dimensional inputs into low-dimensional manifolds have been proposed both as computer vision algorithms and as human vision models (e.g., [25],[41],[71],[73],[80]). In terms of neural substrates, object categorization in the ventral pathway of the primate visual system has been modeled by using the concept of linearly-separable manifolds (rev. [22]). While these remarks highlight the plausibility of manifolds as geometrical representations for complex inputs, much needs to be done to develop and test a comprehensive theory based on manifold representations.
In the following, we will analyze critical paradigms that shed light on the limitations of retinotopic representations and suggest a non-retinotopic manifold representation where motion information and reference frames play an essential role.
V. Insights on Non-Retinotopic Representations from Studies of anorthoscopic perception
A second theory of anorthoscopic perception, “time-of-arrival theory”, posits that a “post-retinal mechanism” stores in memory the information available through the slit and reconstructs a spatially extended version by converting the time-of-arrival into a spatial code [60]. However, this theory has also been refuted based on empirical data. Consider the stimulus shown in Fig. 7 [43],[75],[76]. The stimulus consists of dots depicting two triangular shapes that move in opposite directions. The tips of the triangles pass through the slit simultaneously, followed by the middle segments and finally the base segments. Denote by t0, t1, and t2 with t0<t1<t2 the time instants at which the tip, the middle, and the base of the triangles, respectively, become visible inside the slit. According to the time-of-arrival coding theory, the arrival times for the different parts of the figure t0, t1, and t2 are converted to corresponding spatial positions s0, s1, and s2 with s0<s1<s2. As a result, the theory predicts that observers should perceive the two triangles pointing in the same direction [52]. However, observers perceive the orientation of the two triangles correctly, i.e., the upper triangle pointing to the left and the lower triangle to the right [43],[75],[76]. Thus, these experiments provide strong evidence against the time-of-arrival theory1. A critical observation from these experiments is that, if one does not take into account the direction of motion, then the interpretation of the stimulus will be ambiguous in that mirror-symmetric images moving in opposite directions will generate identical patterns in the slit. Therefore, the direction of motion must be an essential component in building non-retinotopic representations associated with anorthoscopic perception [5],[6],[52],[57].
Based on these observations, we hypothesized that non-retinotopic representations are built using motion vectors and that the metric of non-retinotopic representations depends on the motion vectors. To test these hypotheses, we studied a well-known effect in anorthoscopic perception, viz., compression of perceived form along the direction of motion [1],[28],[32],[43],[46],[60],[66],[92]. A prediction of our hypotheses is that perceived shape should depend on perceived motion as the stimulus moves behind the slit and compression of form should occur when the trailing edge of the stimulus is perceived to move faster than its leading edge. We presented ellipses that moved behind a narrow slit and observers judged the perceived speed of the leading and trailing edges of the stimuli [5]. Fig. 8 shows the data averaged across observers. As predicted, the trailing edge of the ellipse appeared to move faster than the leading edge. In separate sessions, observers judged the compression of the ellipses and the data indicate that, perceptually, the ellipses appeared compressed. We attribute this compression to a velocity-dependent metric used in non-retinotopic representations. As we will discuss below, we suggest a general approach to non-retinotopic representations where an early segmentation of the scene based on motion vectors leads into the construction of local manifolds as the fundamental units of non-retinotopic representations.
VI. Insights on Non-Retinotopic Representations from Studies of Non-retinotopic Feature Perception
Several researchers investigating the metacontrast phenomenon noticed and reported that, even though the mask can render the target invisible, some features of the target could still be seen. Remarkably, these features were not perceived at their retinotopic location but rather were perceptually displaced to appear as part of the mask [33],[34],[78],[79],[82],[87],[88]. This non-retinotopic perception of features was named “feature transposition” [88] or “feature inheritance” [33]. They were typically conceived as errors of the visual system along with other illusions of feature “mis-binding” attributed to lack of attention [83], masking [33],[87],[88], object substitution [26], crowding [59], pooling [8],[59], sampling [16], distributed micro-consciousness [91], and latency differences in stimulus transmission and processing [4],[9]. Yet, several lines of evidence suggest that not all retinotopic mislocalizations reflect errors of visual processing; rather, some follow strategies deployed by the visual system in order to cope with the dynamic constraints outlined in Section II.
It has been shown that when an object is in motion, the information about its shape [53],[54],[55], luminance [74], color [51], and size [40] is integrated along its motion path. For example, Fig. 10 shows a stimulus configuration used by Shimozaki et al. [74]. The first frame contains two squares, one with low luminance and one with high luminance. In frame 2, these squares are shifted downward and randomly either to the right or left creating the percept of downward-right/left motion. In this example, the luminance of the square on the left remained low while the square on the right reduced its luminance to a medium level. Observers were asked to judge the perceived luminance of the target square (see Fig. 10) with respect to the other square in the second frame. Results indicate that luminance (more precisely relative contrast) is integrated along the motion pathway. It is well known that a rapid succession of red and green stimuli at the same retinotopic location leads to the integration of these colors into the percept of yellow [89]. In a recent study, however, it has been shown that the key in this color integration process is not retinotopy but rather the motion pathway of the stimulus (Fig. 11) [51] or selective attention tracking the motion pathway of the stimulus [17]. By using a repetitive version of metacontrast masking (called sequential metacontrast), we have shown that form information is integrated along the motion pathway even when the stimulus containing the form information is invisible (Fig. 12) [54].
In order to pit directly retinotopic against non-retinotopic integration, we used a stimulus known as the “Ternus-Pikler” display [58],[61],[62],[81]. Fig. 13 shows the basic Ternus-Pikler stimulus. The first frame contains three bars that are shifted to the right by one inter-bar distance in the second frame. When the Inter-Stimulus Interval (ISI) between the two frames is brief, observers perceive the leftmost bar in the first frame to move to the rightmost bar in the second frame. The middle two bars appear stationary. This percept is called element motion. When ISI is long, a completely different percept emerges: Now the three bars appear to move together as a group rightwards, a percept called group motion. In order to investigate whether form is processed according to retinotopic or non-retinotopic geometry, we replaced the bars by line segments [53]. The central line segment contained a small offset, known as Vernier offset. Observers were asked to judge the Vernier offset (left of right) of one of the lines in the second frame. In the first experiment, physically none of the lines in the second frame contained an offset. If the geometry of feature integration were retinotopic, the Vernier offset in the middle segment of the first frame would be integrated with the leftmost element in the second frame because these two elements are retinotopically superimposed. As a result, the accordance of the observers’ response with the physical offset of the Vernier in the first frame should be high for the first element in the second frame and should be approximately at chance level for the other two elements. This prediction holds for all ISIs that fall within the temporal integration limit of the human visual system. On the other hand, if the geometry of feature integration is non-retinotopic and follows the rules of motion grouping, different predictions emerge as a function of ISI. As shown in Fig. 13, when ISI is brief, element motion is perceived and the central element of the first frame is perceptually mapped onto the leftmost element in the second frame. In this case, the prediction is identical to that of a retinotopic representation. On the other hand, when ISI is long, according to the group motion percept the central element of the first framed is perceptually mapped onto the central element in the second frame. Thus in this case retinotopic and non-retinotopic predictions are pitted against each other. According to the former, the accordance of the observers’ responses with the Vernier offset in the first frame should be highest for the leftmost element in the second frame. According to the latter, it should be highest for the central element in the second frame. Our data, shown in Fig. 14, are in agreement with the predictions of non-retinotopic motion-grouping based predictions. Several additional experiments further supported the conclusion that the form of moving stimuli is processed according to a non-retinotopic geometry [53].
Recently, we have shown that the Ternus-Pikler display can be used as a simple “litmus test” to investigate whether any given visual process occurs in retinotopic or non-retinotopic representations [10]. The three elements in the Ternus-Pikler display can assume any geometrical shape (e.g., disks) and inside these elements one can insert stimuli of interest. As an example, Fig. 15 shows a version of the Ternus-Pikler display where the stimulus of interest is a visual search display. The search stimuli in the display are defined on two feature dimensions (orientation: vertical/horizontal and color: red/green). The task of the observer is to report whether a target defined by a specific conjunction of two features (e.g., horizontal red bar) is present within the central disk. In conjunction search, performance, as measured by percent correct or reaction times, drops as the number of distractors is increased. According to current models of visual search and attention, each feature is represented in a retinotopic map and attention is required to look up at a given retinotopic location in each map and to bind features of the object at that retinotopic location. In sharp contrast to the predictions of these theories, we found that performance was close to 80% correct in the case of group motion and was near chance when the percept of motion was obliterated [10]. A consistent pattern was observed in reaction-time measures as well [10]. These findings indicate that conjunction search is carried out non-retinotopically. Therefore models of search and attention need to be modified to incorporate non-retinotopic geometries. All the findings reviewed in this and the previous section indicate that non-retinotopic information representation and processing depend critically on motion information. The Ternus-Pikler litmus test demonstrates that motion establishes a reference frame according to which non-retinotopic computations take place. In the next section, we discuss reference frames and the resulting relativity in processes associated with these reference frames.
VII. Relativity in perception and reference frames
The sensitivity of the visual system to relative physical quantities is well documented. For example, as shown by the Ebinghaus-Titchener illusion, a disk surrounded by large disks appears smaller compared to the same size disk surrounded by smaller disks (Fig. 16). An interpretation of this and a variety of other relativity effects is that some of the components in the stimulus serve as a reference frame against which the others are judged. There is evidence that the visual system may be using not a single but a variety of reference frames depending on the stimuli and the task. For example, as the eyes move, stationary stimuli move in retinotopic coordinates while they remain stationary in our perception. It has been therefore suggested that, at higher levels of the visual system, representations that make use of eye-movement command signals are in operation. The geometry of these representations may be spatiotopic in that the reference frames are in spatial rather than retinotopic coordinates.
The importance of reference frames can be further appreciated by considering multi-sensory or sensory-motor coordination. The nervous system is confronted with establishing relations between different sensory-based representations so as to achieve multi-sensory coordination (e.g., audio-visual coordination so that sounds are perceived coherently with their visual sources). In addition, sensory representations need to be converted to motor representations to coordinate sensory-motor tasks (e.g., head or body centered reference frames may be used for motor systems). At the neural level, there is evidence of multiple reference frames including representations that reflect mixture of multiple reference frames [47].
At the input and output levels, the physics of the sensors and actuators, respectively, impose strong constraints that shape neural representations carrying out early encoding of stimuli and the control of actuators. As discussed earlier, the optics of the eye promotes early retinotopic representations. Similarly, the physics and physiology of the plants and the immediate controllers (such as the eye ball and eye muscles) promote pre-motor and motor representations. The major challenge is to understand intermediate representations and geometries that the visual system uses not only to achieve sensorimotor control but also to build a cognitive representation of the world. To study this problem of intermediate representations, we focus here on the representations in the perceptual system that succeed retinotopic representations but that are not necessarily linked to motor representations. A popular conceptual paradigm for such representations is the “object file” approach.
A. Objects as reference frames for visual perception
In his pioneering work, the Gestalt psychologist Joseph Ternus introduced the concept of phenomenal identity: “Experience consists far less in haphazard multiplicity than in the temporal sequence of self-identical objects. We see a moving object, and we say that “this object moves” even though our retinal images are changing at each instant of time and for each place which it occupies in space. Phenomenally the object retains its identity.” [81]. The notion of an object as a representational unit, or a reference frame, appeared in the literature various forms. Kahneman and colleagues introduced the metaphor “object file” [38]. According to this approach, for each object, a file is opened and accessed by the instantaneous location of the object. The visual system collects the sensory information received about the object at that location and inserts it in the object file. For example, Shimozaki et al. interpreted their results showing integration of luminance along the motion path (discussed in Section VI) as evidence for an object-file based (non-retinotopic) representation. Given that there is also evidence that neurons in the higher levels of the ventral pathway exhibit invariant representations that could be labeled as “objects”, the object file metaphor has been a popular approach to conceptualize non-retinotopic representations. While intuitively appealing, this approach has several challenges.
It is difficult to define a priori what an object is. Kahneman et al. acknowledge partly this problem when they discuss hierarchical organization of objects (e.g., the wheel of a car can be an object itself or can be part of the larger object “car” or can be part of the larger object “traffic on the highway”, etc.) and assumed that attention, “bottom-up constraints” and “grouping factors” would define the “preferred level” at which an object file is created. However, the details of these factors and how they determine the proper object level remain unsolved. Indeed, many of Ternus’ original experiments demonstrate the difficulty of deciding a priori which stimulus elements will constitute an object (see for example Fig. 17). As discussed in Section VI, the identities of “objects” can also vary depending on stimulus parameters such as ISI. When ISI is short, the middle two elements constitute one object while when ISI is long, the three elements constitute a single object.
Furthermore, the object-file approach is prone to be circular: to define an object and open a file, the visual system needs to distinguish the object from the background and from other objects in the image. However, this figure-ground segregation for the object necessitates that sufficient information about the object is already available to the visual system. But if the file is not open yet, how can this information be available for analysis? At the computational level, this circularity is a problem of bootstrapping. As we have argued in Section II, the read-out and processing of information for moving objects cannot be achieved in a satisfactory manner in retinotopic representations (“moving ghosts problem”) and therefore a non-retinotopic representation needs to be created before even an object file can be opened (if such a thing exists). In the next section, we discuss what this non-retinotopic representation can be.
B. Local manifolds induced by motion segregation
An alternative approach in building hierarchical representations relies on lower level motion analysis as opposed to high level object analysis. Johansson has been one of the pioneers in this area. From a perceptual geometrical perspective, he replaced Euclidian geometry by perspective geometry and devised several studies showing how motion can be analyzed in hierarchical frames of reference within perspective geometrical representations [35][36][37]. He introduced powerful demonstrations wherein complex biological motion, such as a walking human, is sampled by few visible points placed on joints, arms, legs, torso, etc. While the individual motion paths of these sampled points appear complex and unrelated, observers are able to organize them into a hierarchy of meaningful motion patterns that reveal the underlying biological motion. Johansson also proposed a vector analysis method to model his data [35]. Although the proposed vector analysis method has limitations [42], it serves as a good heuristic as a starting point.
To gain insights into geometric representations underlying dynamic form perception, let us consider the “moving ghosts” problem. In order to integrate features of a moving object, the ideal approach would be to compensate for its motion so that it can be stabilized at a locus where integration occurs (e.g., [1],[56]). In normal viewing, pursuit eye movements partly accomplish this goal for a selected target. However, the environment contains multiple objects moving with different velocities and thus an eye movement or a single global motion-compensation scheme cannot “stabilize” all stimuli. We suggest that the visual system determines locally common motion vectors that can be used to minimize motion variations within a local neighborhood. The heuristic behind this local approach is that, due to their contiguous spatial extent, objects map into local neighborhoods in the retinotopic space. When an object moves, its various parts share a common motion vector, even though different parts of the object may have different velocities (e.g., the tail of a panther may be swinging in addition to the lateral motion of the animal). The “common motion vectors” in local retinotopic neighborhoods can be used to partially stabilize moving objects. The use of common motion in image grouping and segmentation (e.g., Gestalt law of “common faith”), common motion vector decomposition and the attendant relativity of motion have extensive empirical support [21],[35],[36],[39],[42],[65],[84],[85] and is related to the Johansson’s motion vector analysis. However, the exact rules of common vector determination and the nature of reference frames are poorly understood. Notwithstanding the details of how they are computed, one can view these common motion vectors as local reference frames to allow relative motion and form computations at non-retinotopic loci.
However, this local motion vector approach cannot be sufficient by itself to stabilize stimuli because (i) as mentioned above, objects can have multiple parts moving with different velocities, (ii) as an object moves in space and different parts of the same object become gradually exposed to the observer, the perspective projection of the object and therefore the underlying motion vectors can vary, (iii) velocities can vary for non-rigid objects/parts, (iv) at this early stage of processing, the visual system has not yet segregated an object from neighboring objects nor the background. The points (i)–(iii) imply that stabilization will not be perfect and the non-retinotopic representation should be able to accommodate residual motion vectors and be able to represent shape transformations that a moving object can undergo. In order to meet these constraints, we suggest that the non-retinotopic representations will have properties akin to manifolds of the mathematical field of topology. In other words, while the representation preserves local neighborhood relations, it is “stretchable” (cf. rubber-sheet geometry). The point (iv) implies that the units in non-retinotopic representations are not “objects”. Thus instead of concepts such as “object files” [38], we propose that non-retinotopic manifolds represent local geometries around reference frames in order to enable the visual system to accrue and process form information. The processed form information may belong to one or multiple objects. The notions of topology and non-Euclidian geometries have been previously used in visual perception [15],[18],[19],[29],[30],[48]. In our use, these geometric concepts represent the transformations from retinotopic to non-retinotopic representations for dynamic form computation. The proposed approach is depicted in Fig. 18. At the bottom of the figure, in the retinotopic space, a group of dots move rightwards (highlighted in red) while a group of dots move upwards (highlighted in orange). Based on differences in motion vectors, the two local neighborhoods are mapped into two different non-retinotopic representations; for clarity the figure shows only the non-retinotopic representation for rightward moving dots. A common vector for the neighborhood is determined (dashed green vector) and serves as the reference frame for the neighborhood. Accordingly, all motion vectors are decomposed into a sum of the reference motion and a residual motion vector. The stimulus in the local neighborhood is mapped on a manifold (in Fig. 18, for depiction purposes a sphere is used), i.e. a geometric structure that preserves local neighborhood relations. However, the surface can be stretched and deformed. The residual motion vectors, or relative motion components with respect to the reference frame, are then applied to the manifold so as to deform it to induce transformations that the shape undergoes during motion. Although these ideas need extensive empirical tests and refinements, the existing data discussed herein provide a good basis as a starting point. Ternus-Pikler displays indicate that a motion-based grouping is fundamental in non-retinotopic feature processing. Similarly, in a recent study we have shown that non-retinotopic processing observed in metacontrast is strongly correlated with perceived motion indicating that motion is the carrier for non-retinotopic feature processing [11]. As discussed in our review of anorthoscopic perception, which provides the strongest example of form perception in the absence of retinotopic representation, motion is essential in setting non-retinotopic representations. Furthermore, our studies on distortion effects support the view that motion vectors “stretch” non-retinotopic manifolds.
VIII. Conclusions
Understanding how information is represented and processed in the brain is key to reverse-engineering perceptual, cognitive, and motor functions carried out by the nervous system. Vision is the most predominant sense in humans and the visual system is the most extensively studied part of the cortex. While it is clear that geometry and visual perception are closely linked, very little is known about the geometry of the visual system beyond its early retinotopic organization. In this paper, we reviewed several lines of evidence that show the insufficiency of retinotopic representations in supporting real-time perception. The object-file concept, while intuitively appealing for non-retinotopic representations has several shortcomings. Based on our recent studies as well as findings published in the literature, we proposed an alternative approach where the units of non-retinotopic representations are not objects but rather local manifolds created by motion segmentation. We also proposed that the metric of these manifolds is based on relative motion vectors. While the proposed approach awaits further investigations, we believe that it will be a fruitful avenue combining the concepts of relativity and non-Euclidian geometries in deciphering the geometry of visual perception and the underlying neural representations.
Acknowledgments
This work was supported in part by the National Institutes of Health (NIH) under Grant R01 EY018165.
Footnotes
Note that the retinal painting theory also makes the same prediction as the time-of-arrival theory. Because the eye movement is common to all parts of the stimulus, the same retinotopic image will be painted for the upper and lower triangles. Thus, the aforementioned experiments also provide additional evidence against the retinal painting theory.
The content is solely the responsibility of the authors and does not necessarily represent the official views of NIH.
Contributor Information
Haluk Öğmen, Email: ogmen@uh.edu, Department of Electrical & Computer Engineering and Center for NeuroEngineering & Cognitive Science, University of Houston, Houston, TX 77204-4005 USA (phone: 713-743-4428; fax: 713-743-4444.
Michael H. Herzog, Email: michael.herzog@epfl.ch, Brain Mind Institute, Ecole Polytechnique Fédérale de Lausanne (EPFL), CH-1015 Lausanne, Switzerland
References
- 1.Anderson CH, van Essen DC. Shifter circuits: A computational strategy for dynamic aspects of visual processing. Proceedings of the National Academy of Science, USA. 1987;84:6297–6301. doi: 10.1073/pnas.84.17.6297. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Anstis SM, Atkinson J. Distortions in moving figures viewed through a stationary slit. American Journal of Psychology. 1967;80:572–585. [PubMed] [Google Scholar]
- 3.Arcaro MJ, McMains SA, Singer BD, Kastner S. Retinotopic organization of human ventral visual cortex. J of Neuroscience. 2009;29:10638–10652. doi: 10.1523/JNEUROSCI.2807-09.2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Arnold DH, Clifford CW. Determinants of asynchronous processing in vision. Proc R Soc Lond B. 2002;269:579–583. doi: 10.1098/rspb.2001.1913. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Aydın M, Herzog MH, Öğmen H. Perceived Speed Differences Explain Apparent Compression in Slit Viewing. Vision Research. 2008;48:1603–1612. doi: 10.1016/j.visres.2008.04.020. [DOI] [PubMed] [Google Scholar]
- 6.Aydın M, Herzog MH, Öğmen H. Shape distortions and Gestalt grouping in anorthoscopic perception. Journal of Vision. 2009;9(3):8, 1–8. doi: 10.1167/9.3.8. http://journalofvision.org/9/3/8/ [DOI] [PMC free article] [PubMed]
- 7.Bachmann T. Psychophysiology of Visual Masking: The Fine Structure of Conscious Experience. New York: Nova Science Publishers; 1994. [Google Scholar]
- 8.Baldassi S, Burr DC. Feature-based integration of orientation signals in visual search. Vision Research. 2000;40:1293–1300. doi: 10.1016/s0042-6989(00)00029-8. [DOI] [PubMed] [Google Scholar]
- 9.Bedell HE, Chung STL, Öğmen H, Patel SS. Color and motion: which is the tortoise and which is the hare? Vision Research. 2003;43:2403–2412. doi: 10.1016/s0042-6989(03)00436-x. [DOI] [PubMed] [Google Scholar]
- 10.Boi M, Öğmen H, Krummenacher J, Otto T, Herzog MH. A (fascinating) litmus test for human retino- vs. non-retinotopic processing. Journal of Vision. 2009;9(13):5, 1–11. doi: 10.1167/9.13.5. http://journalofvision.org/9/13/5/ [DOI] [PMC free article] [PubMed]
- 11.Breitmeyer BG, Herzog MH, Öğmen H. Motion, not masking, provides the medium for feature attribution. Psychological Science. 2008;19:823–829. doi: 10.1111/j.1467-9280.2008.02163.x. [DOI] [PubMed] [Google Scholar]
- 12.Breitmeyer BG, Öğmen H. Visual Masking: Time Slices through Conscious and Unconscious Vision. 2. Oxford University Press; Oxford, UK: 2006. [Google Scholar]
- 13.Brewer AA, Press WA, Logothetis NK, Wandell BA. Visual areas in macaque cortex measured using functional magnetic resonance imaging. J Neurosci. 2002;22:10416–10426. doi: 10.1523/JNEUROSCI.22-23-10416.2002. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Butler BE, Mewhort DH, Browse RA. When do letter features migrate? A boundary condition for feature-integration theory. Perception & Psychophysics. 1991;49:91–99. doi: 10.3758/bf03211620. [DOI] [PubMed] [Google Scholar]
- 15.Caelli T, Hoffman W, Lindman H. Subjective Lorenz transformations and the perception of motion. J Opt Soc Am. 1978;68:402–411. doi: 10.1364/josa.68.000402. [DOI] [PubMed] [Google Scholar]
- 16.Cai R, Schlag J. A new form of illusory conjunction between color and shape. Journal of Vision. 2001;1:127a. [Google Scholar]
- 17.Cavanagh P, Holcombe AO, Chou W. Mobile computation: Spatiotemporal integration of the properties of objects in motion. Journal of Vision. 2008;8(12):1, 1–23. doi: 10.1167/8.12.1. http://journalofvision.org/8/12/1/ [DOI] [PMC free article] [PubMed]
- 18.Chen L. Topological structure in visual perception. Science. 1982;218:699–700. doi: 10.1126/science.7134969. [DOI] [PubMed] [Google Scholar]
- 19.Chen L. The topological approach to perceptual organization. Visual Cognition. 2005;12:553–637. [Google Scholar]
- 20.Chen S, Bedell HE, Öğmen H. A target in real motion appears blurred in the absence of other proximal moving targets. Vision Research. 1995;35:2315–2328. doi: 10.1016/0042-6989(94)00308-9. [DOI] [PubMed] [Google Scholar]
- 21.Cutting JE, Proffitt DR. The minimum principle and the perception of absolute, common, and relative motions. Cognitive Psychology. 1982;14:211–246. doi: 10.1016/0010-0285(82)90009-3. [DOI] [PubMed] [Google Scholar]
- 22.DiCarlo JJ, Cox DD. Untangling invariant object recognition. Trends in Cognitive Sciences. 2007;11:333–341. doi: 10.1016/j.tics.2007.06.010. [DOI] [PubMed] [Google Scholar]
- 23.Dougherty RF, Koch VM, Brewer AA, Fischer B, Modersitzki J, Wandell BA. Visual field representations and locations of visual areas V1/2/3 in human visual cortex. Journal of Vision. 2003;3(10):586–598. doi: 10.1167/3.10.1. [DOI] [PubMed] [Google Scholar]
- 24.Duda RO, Hart PE, Stork DG. Pattern Classification. 2. Wiley; New York: 2001. [Google Scholar]
- 25.Elgammal A, Lee CS. Nonlinear manifold learning for dynamic shape and dynamic appearance. Computer Vision and Image Understanding. 2007;106:31–46. [Google Scholar]
- 26.Enns J. Visual binding in the standing wave illusion. Psychonomic Bulletin & Review. 2002;9:489–496. doi: 10.3758/bf03196304. [DOI] [PubMed] [Google Scholar]
- 27.Fendrich R, Rieger JW, Heinze HJ. The effect of retinal stabilization on anorthoscopic percepts under free-viewing conditions. Vision Research. 2005;45:567–582. doi: 10.1016/j.visres.2004.09.025. [DOI] [PubMed] [Google Scholar]
- 28.Haber RN, Nathanson LS. Post-retinal storage? Some further observations on Parks’ camel as seen through the eye of a needle. Perception & Psychophysics. 1968;3:349–355. [Google Scholar]
- 29.He S. Holes, objects, and the left hemisphere. Proceedings of the National Academy of Sciences. 2008;105:1103–1104. doi: 10.1073/pnas.0710631105. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Hecht H, Bader H. Perceiving topological structure of 2-D patterns. Acta Psychologica. 1998;99:255–292. doi: 10.1016/s0001-6918(98)00015-8. [DOI] [PubMed] [Google Scholar]
- 31.Heeger D. Perception Lecture Notes. New York University; 2006. http://www.cns.nyu.edu/~david/courses/perception/lecturenotes/ [Google Scholar]
- 32.von Helmholtz H. Handbook of Physiological Optics. New York: Dover Reprint; 1867. 1963. [Google Scholar]
- 33.Herzog MH, Koch C. Seeing properties of an invisible object: Feature inheritance and shine-through. Proc Natl Acad Sci USA. 2001;98:4271–4275. doi: 10.1073/pnas.071047498. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Hofer D, Walder F, Groner M. „Metakontrast: ein berühmtes, aber schwer messbares Phänomen. Schweiz Zeitschr Psychol. 1989;48:219–232. [Google Scholar]
- 35.Johansson G. Visual perception of biological motion and a model for its analysis. Perception & Psychophysics. 1973;14:201–211. [Google Scholar]
- 36.Johansson G. Visual motion perception. Scientific American. 1975;232:76–88. doi: 10.1038/scientificamerican0675-76. [DOI] [PubMed] [Google Scholar]
- 37.Johansson G. Spatio-temporal differentiation and Integration in visual motion perception: An experimental and theoretical analysis of calculus-like functions in visual data processing. Psychol Res. 1976;38:379–393. doi: 10.1007/BF00309043. [DOI] [PubMed] [Google Scholar]
- 38.Kahneman D, Treisman A, Gibbs BJ. The reviewing of object files: Object-specific integration of information. Cognitive Psychology. 1992;24:175–219. doi: 10.1016/0010-0285(92)90007-o. [DOI] [PubMed] [Google Scholar]
- 39.Kalveram KT, Ritter M. The formation of reference systems in visual motion perception. Psychol Res. 1979;40:397–405. doi: 10.1007/BF00309419. [DOI] [PubMed] [Google Scholar]
- 40.Kawabe T. Spatiotemporal feature attribution for the perception of visual size. Journal of Vision. 2008;8(8):7, 1–9. doi: 10.1167/8.8.7. http://journalofvision.org/8/8/7/ [DOI] [PubMed]
- 41.Lin T, Zha H. Riemannian manifold learning. IEEE Transactions on Pattern Analysis and Machine intelligence. 2008;30:796–809. doi: 10.1109/TPAMI.2007.70735. [DOI] [PubMed] [Google Scholar]
- 42.Mack A. Perceptual aspects of motion in the frontal plane. In: Boff KR, Kaufman L, Thomas JP, editors. Handbook of Perception and Human Performance. Vol. 1. New York: Wiley; 1986. pp. 17-1–17-38. [Google Scholar]
- 43.McCloskey M, Watkins MJ. The seeing-more-than-is-there phenomenon: Implications for the locus of iconic storage. Journal of Experimental Psychology: Human Perception and Performance. 1978;4:553–564. doi: 10.1037//0096-1523.4.4.553. [DOI] [PubMed] [Google Scholar]
- 44.Michotte A, Thinès G, Crabbé G. Studia Psychologica. Institut de Psychologie de l’Université de Louvain; 1964. Les Compléments Amodaux des Structures Perceptives. [Google Scholar]
- 45.Minsky ML, Papert SA. Perceptrons. MIT Press; Cambridge, MA: 1969. (Expanded edition, 1988) [Google Scholar]
- 46.Morgan MJ, Findlay JM, Watt RJ. Aperture viewing: A review and a synthesis. Quarterly Journal of Experimental Psychology. 1982;34A:211–233. doi: 10.1080/14640748208400837. [DOI] [PubMed] [Google Scholar]
- 47.Mullette-Gillman OA, Cohen YE, Groh JM. Eye-Centered, Head-Centered, and Complex Coding of Visual and Auditory Targets in the Intraparietal Sulcus. J Neurophysiol. 2005;94:2331–2352. doi: 10.1152/jn.00021.2005. [DOI] [PubMed] [Google Scholar]
- 48.Müsseler J. Perceiving and measuring of spatiotemporal events. In: Jordan JS, editor. Modeling consciousness across the disciplines. Maryland: University Press of America, Inc; 1999. pp. 95–112. [Google Scholar]
- 49.Naito S, Cole JB. The Gravity Lens Illusion and its Mathematical Model. In: Fischer GH, Laming D, editors. Contributions to Mathematical Pychology, Psychometrics, and Methodology. New York: Springer Verlag; 1994. [Google Scholar]
- 50.Nishida S. Motion-based analysis of spatial patterns by the human visual system. Current Biology. 2004;14:830–839. doi: 10.1016/j.cub.2004.04.044. [DOI] [PubMed] [Google Scholar]
- 51.Nishida S, Watanabe J, Kuriki I, Tokimoto T. Human brain integrates colour signals along motion trajectory. Current Biology. 2007;17(4):366–372. doi: 10.1016/j.cub.2006.12.041. [DOI] [PubMed] [Google Scholar]
- 52.Öğmen H. A theory of moving form perception: Synergy between masking, perceptual grouping, and motion computation in retinotopic and non-retinotopic representations. Advances in Cognitive Psychology. 2007;3:67–84. doi: 10.2478/v10053-008-0015-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 53.Öğmen H, Otto T, Herzog MH. Perceptual grouping induces non-retinotopic feature attribution in human vision. Vision Research. 2006;46:3234–3242. doi: 10.1016/j.visres.2006.04.007. [DOI] [PubMed] [Google Scholar]
- 54.Otto TU, Öğmen H, Herzog MH. The flight path of the phoenix – The visible trace of invisible elements in human vision. Journal of Vision. 2006;6:1079–1086. doi: 10.1167/6.10.7. [DOI] [PubMed] [Google Scholar]
- 55.Otto TU, Öğmen H, Herzog MH. Assessing the microstructure of motion correspondences with non-retinotopic feature attribution. Journal of Vision. 2008;8(7):16, 1–15. doi: 10.1167/8.7.16. [DOI] [PubMed] [Google Scholar]
- 56.Paakkonen AK, Morgan MJ. Effect of motion on blur discrimination. J Opt Soc Am A. 1994;11:992–1002. [Google Scholar]
- 57.Palmer EM, Kellman PJ, Shipley TF. A theory of dynamic occluded and illusory object perception. Journal of Experimental Psychology: General. 2006;135:513–541. doi: 10.1037/0096-3445.135.4.513. [DOI] [PubMed] [Google Scholar]
- 58.Pantle A, Picciano L. A multistable movement display: Evidence for two separate motion systems in human vision. Science. 1976;193:500–502. doi: 10.1126/science.941023. [DOI] [PubMed] [Google Scholar]
- 59.Parkes L, Lund J, Angelucci A, Solomon JA, Morgan M. Compulsory averaging of crowded orientation signals in human vision. Nature Neuroscience. 2001;4:739–744. doi: 10.1038/89532. [DOI] [PubMed] [Google Scholar]
- 60.Parks TE. Post-retinal visual storage. American Journal of Psychology. 1965;78:145–147. [PubMed] [Google Scholar]
- 61.Petersik JT, Rice CM. The evolution of explanations of a perceptual phenomenon: A case history using the Ternus effect. Perception. doi: 10.1068/p5522. (in press) [DOI] [PubMed] [Google Scholar]
- 62.Pikler J. Sinnesphysiologische Untersuchungen. Leipzig: Barth; 1917. [Google Scholar]
- 63.Plateau J. Bulletin de l’Académie des Sciences et Belles Lettres de Bruxelles. Vol. 3. Bruxelles; 1836. Notice sur l’Anorthoscope; pp. 7–10. [Google Scholar]
- 64.Purushothaman G, Öğmen H, Chen S, Bedell HE. Motion deblurring in a neural network model of retino-cortical dynamics. Vision Research. 1998;38:1827–1842. doi: 10.1016/s0042-6989(97)00350-7. [DOI] [PubMed] [Google Scholar]
- 65.Restle F. Coding theory of the perception of motion configurations. Psychological Review. 1979;89:1–24. [PubMed] [Google Scholar]
- 66.Rock I. Anorthoscopic perception. Scientific American. 1981;244:145–153. doi: 10.1038/scientificamerican0381-145. [DOI] [PubMed] [Google Scholar]
- 67.Rock I, Halper F, DiVita J, Wheeler D. Eye movement as a cue to figure motion in anorthoscopic perception. Journal of Experimental Psychology: Human Perception and Performance. 1987;13:344–352. doi: 10.1037//0096-1523.13.3.344. [DOI] [PubMed] [Google Scholar]
- 68.Rock I, Sigman E. Intelligence factors in the perception of form through a moving slit. Perception. 1973;2:357–369. doi: 10.1068/p020357. [DOI] [PubMed] [Google Scholar]
- 69.Rosenblatt F. Principles of Neurodynamics. Spartan Books; New York: 1962. [Google Scholar]
- 70.Rothschild H. Untersuchungen über die sogenannten anorthoskopischen zerrbilder. Zeitschrift für Psychologie. 1922;90:137–166. [Google Scholar]
- 71.Roweis ST, Saul LK. Nonlinear dimensionalty reduction by locally linear embedding. Science. 2000;290:2323–2326. doi: 10.1126/science.290.5500.2323. [DOI] [PubMed] [Google Scholar]
- 72.Saygin PA, Sereno MI. Retinotopy and attention in human occipital, temporal, parietal, and frontal cortex. Cerebral Cortex. 2008;18:2158–2168. doi: 10.1093/cercor/bhm242. [DOI] [PubMed] [Google Scholar]
- 73.Seung HS, Lee DD. The manifold ways of perception. Science. 2000;290:2268–2269. doi: 10.1126/science.290.5500.2268. [DOI] [PubMed] [Google Scholar]
- 74.Shimozaki SS, Eckstein MP, Thomas JP. The maintenance of apparent luminance of an object. Journal of Experimental Psychology: Human Perception and Performance. 1999;25(5):1433–1453. doi: 10.1037//0096-1523.25.5.1433. [DOI] [PubMed] [Google Scholar]
- 75.Sohmiya T, Sohmiya K. Where does an anorthoscopic image appear? Perceptual and Motor Skills. 1992;75:707–714. doi: 10.2466/pms.1992.75.3.707. [DOI] [PubMed] [Google Scholar]
- 76.Sohmiya T, Sohmiya K. What is a crucial determinant in anorthoscopic perception? Perceptual and Motor Skills. 1994;78:987–998. doi: 10.1177/003151259407800357. [DOI] [PubMed] [Google Scholar]
- 77.Sokolov A, Pavlova M. Visual motion detection in hierarchical spatial frames of reference. Exp Brain Res. 2006;174:477–486. doi: 10.1007/s00221-006-0487-6. [DOI] [PubMed] [Google Scholar]
- 78.Stewart AL, Purcell DG. U-shaped masking functions in visual backward masking: Effects of target configuration and retinal position. Perception & Psychophysics. 1970;7:253–256. [Google Scholar]
- 79.Stoper AE, Banffy S. Relation of split apparent motion to metacontrast. Journal of Experimental Psychology: Human Perception and Performance. 1977;3:258–277. doi: 10.1037//0096-1523.3.2.258. [DOI] [PubMed] [Google Scholar]
- 80.Tenenbaum JB, de Silva V, Langford JC. A global geometric framework for nonlinear dimensionality reduction. Science. 2000;290:2319–2323. doi: 10.1126/science.290.5500.2319. [DOI] [PubMed] [Google Scholar]
- 81.Ternus J. Experimentelle Untersuchung über phänomenale Identität. Psychologische Forschung. 1926;7:81–136. [Google Scholar]
- 82.Toch HH. The perceptual elaboration of stroboscopic presentation. Am J Psychol. 1956;69:345–358. [PubMed] [Google Scholar]
- 83.Treisman A, Schmidt H. Illusory conjunctions in the perception of objects. Cognitive Psychology. 1982;14:107–141. doi: 10.1016/0010-0285(82)90006-8. [DOI] [PubMed] [Google Scholar]
- 84.Wallach H. The perception of motion. Scientific American. 1959;201:56–60. doi: 10.1038/scientificamerican0759-56. [DOI] [PubMed] [Google Scholar]
- 85.Wallach H, Becklen R, Nitzberg D. Vector analysis and process combination in motion perception. Journal of Experimental Psychology: Human Perception and Performance. 1985;11:93–102. doi: 10.1037//0096-1523.11.1.93. [DOI] [PubMed] [Google Scholar]
- 86.Watanabe J, Nishida S. Veridical perception of moving colors by trajectory integration of input signals. Journal of Vision. 2007;7(11):3, 1–16. doi: 10.1167/7.11.3. [DOI] [PubMed] [Google Scholar]
- 87.Werner H. Studies on contour: I. Qualitative analyses. Am J Psychol. 1935;47:40–64. [Google Scholar]
- 88.Wilson AE, Johnson RM. Transposition in backward masking. The case of travelling gap. Vision Research. 1985;25:283–288. doi: 10.1016/0042-6989(85)90120-8. [DOI] [PubMed] [Google Scholar]
- 89.Wyszecki G, Stiles WS. Quantitative Data and Formulae. 2. New York: Wiley; 1982. Color Science: Concepts and Methods. [Google Scholar]
- 90.Yin C, Shimojo S, Moore C, Engel SA. Dynamic shape integration in extrastriate cortex. Current Biology. 2002;12:1379–1385. doi: 10.1016/s0960-9822(02)01071-0. [DOI] [PubMed] [Google Scholar]
- 91.Zeki S. Localization and globalization in conscious vision. Annu Rev Neurosci. 2001;24:57–86. doi: 10.1146/annurev.neuro.24.1.57. [DOI] [PubMed] [Google Scholar]
- 92.Zöllner F. Über eine neue art anorthoskopischer zerrbilder. Annalen der Physik und Chemie. 1862;117:477–484. [Google Scholar]