Skip to main content

This is a preprint.

It has not yet been peer reviewed by a journal.

The National Library of Medicine is running a pilot to include preprints that result from research funded by NIH in PMC and PubMed.

bioRxiv logoLink to bioRxiv
[Preprint]. 2024 Sep 18:2024.09.10.612333. [Version 2] doi: 10.1101/2024.09.10.612333

Anisotropy of object non-rigidity: High-level perceptual consequences of cortical anisotropy

Akihito Maruya 1, Qasim Zaidi 1,*
PMCID: PMC11429613  PMID: 39345500

Abstract

This study presents instances where variations in a complex, higher-level perceptual phenomenon— an anisotropy in object non-rigidity is explained by the distribution of low-level neural properties in the primary visual cortex. Specifically, we examine the visual interpretation of two rigidly connected rotating circular rings. At speeds where observers predominantly perceive rigid rotation of the rings rotating horizontally, observers perceive only non-rigid wobbling when the image is rotated by 90°. Additionally, vertically rotating rings appear narrower and longer compared to their physically identical horizontally rotating counterparts. We show that these perceived shape changes can be decoded from V1 outputs by considering anisotropies in orientation-selective cells. We then empirically demonstrate that even when the vertically rotating ellipses are widened or the horizontally rotating ellipses are elongated so that the shapes match, the perceived difference in non-rigidity is reduced only by a small amount and increased non-rigidity persists in vertical rotations, suggesting that motion mechanisms also play a role. By incorporating cortical anisotropies into optic flow computations, we show that motion gradients for vertical rotations align more with physical non-rigidity, while horizontal rotations align closer to rigidity, indicating that cortical anisotropies contribute to the heightened perception of non-rigidity when orientation shifts from horizontal to vertical. The study underscores the importance of low-level anisotropies in shaping high-level percepts and raises questions about their evolutionary significance, particularly for shape constancy and motion perception.

Introduction

We present a remarkable instance of variations in a complex higher-level percept explained directly from the distribution of low-level neural properties in primary visual cortex by using a combination of mathematical derivations and computational simulations to reproduce the results of psychophysical experiments. The higher-level percept we demonstrate is an unexpected anisotropy in object non-rigidity, and the low-level properties are documented cortical anisotropies in numbers and tuning widths of orientation selective and direction selective neurons. The video in Figure 1A shows two connected Styrofoam rings on a rotating turntable. Most observers report that the two rings are rigidly joined and rotating together, although sometimes the top ring is reported as moving independently from the bottom ring and rolling or wobbling, seemingly defying physical plausibility. In the video in Figure 1B, the right ring is almost always reported as moving independently, wobbling against the other ring. The context reveals that one video is just a 90° rotation of the other, so clearly an explanation is needed for the increase in nonrigidity.

Figure 1. Anisotropy of a rotating ring illusion videos.

Figure 1.

A: Two Styrofoam rings glued together at an angle are seen to rotate together on the turntable. B: When rotated 90°, one ring appears to move independently and wobble against the ring on the turntable. C & D: Two circular (C) and octagonal (D) rings physically rotate together with a fixed connection at the junction. E & F: The wo pairs of rings physically move independently and wobble against each other. For the circular rings, E is indistinguishable from C. However, with trackable vertices, F is discernable from D. G & H are the same as C & D except that they are rotated 90° and they both look non-rigidly connected.

3D-shape-from-motion models typically assume that objects are rigid, which simplifies the math (Tomasi & Kanade, 1992) but organisms generally change shape to move, eat, observe, mate, etc., so articulated and elastic motions are common in the real world (Aggarwal et al., 1998), and human observers are as good at discerning 3D shapes of non-rigid objects as of rigid objects (Jain & Zaidi, 2011). Maruya and Zaidi (2024) showed that a rotating rigid object with two salient parts, such as the object in Fig 1, can appear rigid or non-rigid depending on speed of rotation and the presence of salient features. At moderate to high speeds, the nonrigidity is predictable from the motion-energy flow, whereas at slow speeds, feature tracking dominates (Lu & Sperling, 2001; Zaidi & DeBonet, 2000), if trackable features are present (Shiffrar & Pavel, 1991; Lorenceau & Shiffrar, 1992; Mach, 1886; Weiss & Adelson, 2000; Rokers, Yuille, & Liu, 2006; Vezzani, Kramer, & Bressan, 2014). The change in non-rigidity when the image is rotated by 90° could be due to anisotropy in shape or motion mechanisms. We investigate both as possible consequences of documented anisotropies in the population distribution of orientation and direction selective neurons.

To function successfully in the objective reality of the physical world, an organism’s perceptions must lead to accurate inferences of attributes such as shapes, sizes, locations, distances, motions, colors, textures, and kinds (Burge, 2000). The earliest discovered cave drawings (Oktaviana et al., 2024) suggest that shape is the most important attribute for humans to identify objects and animals. In mathematics, the shape of a rigid object is a geometric attribute that remains unchanged across variations in location, rotation, and scale (Kendall et al., 1999), which corresponds to viewpoint invariance for a visual system. The shape of an object would not be a useful attribute if the mathematical definition was violated when an object was viewed from different angles. Unfortunately shape constancy is often violated. For example, horizontal parallelogram balconies on a tall building are seen as rectangles tilted up or down depending on viewpoint (Griffiths and Zaidi, 2000). Such violations have been attributed to a tendency to assume right angles (Griffiths and Zaidi, 2000; Dvoeglazova & Sawada, 2024), to try to maximize compactness or symmetry (Pizlo & De Barros, 2021), but the neural substrates of such processes are completely unknown. A more understandable kind of shape-constancy violation for 3D objects is caused by simply rotating the object in the frontal-plane. Cohen and Zaidi (Cohen & Zaidi, 2007) showed that the perceived depth of a wedge defined by shape-from-texture cues is a function of the orientation of the shape. They linked this to a 2D anisotropy in the perception of an angle as a function of its orientation and showed that the 2D effect could be explained by a model based on the documented anisotropy in number and selectivity of orientation tuned cells in primary visual cortex of mammals which show the greatest concentration and narrower tuning for horizontal orientations, then vertical, and then oblique (Li et al., 2003) The anisotropy in primary visual cortex has been linked to encoding the statistical distribution of orientations in the world (Girshick et al., 2011; Harrison et al., 2023), thus suggesting a common cause for the perceived effects of object orientation and the “oblique effect” where observers exhibit superior performance for stimuli presented along the cardinal (vertical and horizontal) axes as compared to oblique axes (Appelle, 1972). The cortical anisotropy has now been well documented for primate brains (Fang et al., 2022), where the anisotropy seems to increase from V1 to V4, raising the question of why evolution leads to brain properties that give more importance to the detection/discrimination of isolated orientations than to preserving perceived shapes across image rotations.

The neural bases of anisotropy of motion processing and perception is less well understood. Psychophysical anisotropies favoring cardinal directions over oblique directions have been shown for direction-discrimination but not for direction-detection of translating stimuli (Coletta et al., 1993; Gros et al., 1998; Dakin et al., 2005; Steven et al., 2005; Rokem & Silver, 2009). The anisotropies that have been documented in area MT (Maunsell and Van, 1983; Churchland et al., 2003) seem to not be systematic, although the number of cells measured is much less than in the V1 studies. There are some reports of direction selectivity for object motion in V4 (Bigelow et al., 2023), where there are documented anisotropies (Fang et al., 2022), but the two issues have not been studied jointly. Given that most V1 simple cells receiving Magnocellular input from LGN are direction selective (Chariker et al., 2022), in our modeling, we assume that the direction preference and tuning of cells in mammalian striate cortex could be assumed to correspond to orientation preference along the axis orthogonal to the motion direction (Li et al., 2003). To our knowledge, anisotropies in motion based higher level perception, such as 3D-shape-from-motion or rigidity versus nonrigidity, have not been studied previously.

Shape and non-rigidity anisotropies are also easy to see when we switch from physical objects to computer graphics and orient the object to be symmetrical around the vertical or horizontal axis, such as the videos of the ring-pairs in Figures 1C1G. Figure 1C & 1D simulate rigidly rotating configurations and Figures 1E & 1F simulate rings that wobble and slide non-rigidly against each other. When attention is directed to the connecting joint, rigidity and non-rigidity are both easy to see in the octagonal rings because the vertices at the junction of the two rings either stay together or slide past each other. The circular rings, however, can be seen as rigidly rotating or non-rigidly wobbling, because without features, rigidly rotating and wobbling rings generate identical retinal images. Consequently, whether they are seen as rigidly rotating or non-rigidly wobbling depends on how the images are processed by the visual system. If the image is rotated by 90 degrees, the non-rigidity becomes much more pronounced for both types of rings (Figure 1G & 1H). Figures in the 2nd and 4th rows show a single frame of horizontally or vertically rotating rings, where the intersection of the two rings faces the viewer. One noticeable difference is that the vertically elongated rings appear narrower and longer than the horizontal ones. By turning the page 90 degrees, the reader can ascertain that this shape illusion is a function of object orientation.

Are the shape changes responsible for the increased nonrigidity, or is it due to a cortical anisotropy in velocity sensitive neurons? We first show that these perceived shape changes can be decoded from V1 outputs, by considering anisotropies in the number and tuning widths of orientation-selective cells. Then we show empirically that when observers widened the vertically rotating ellipses or elongated the horizontally rotating ellipses so that the shapes matched, the perceived difference in nonrigidity was reduced but increased nonrigidity was still perceived for the vertical rotation. We next incorporated cortical anisotropies into motion flow computations. The estimated motion fields were decomposed into gradients of divergence, curl, and deformation. Rigidity of rotating 2D shapes can be tested by showing that only the curl component is non-zero (Todorović,1993), but this is not true for the projections of the 3D rings, so we had to compare all three gradients generated by physical rotation versus physical wobbling. The gradients for vertical rotation of the rings more closely matched physical wobbling, while the gradients for horizontal rotation fell between physical wobbling and rotation, suggesting that hardwired cortical anisotropies can explain the increase in perceived non-rigidity from horizontal to vertical rotation.

Results

We will first test whether the static shape anisotropy can account for the nonrigidity anisotropy, and if that does not account for most of the illusion, we will then examine velocity anisotropy.

Shape Anisotropy

Figures 2A & 2B present the two orientations of the rings in the phase that best demonstrates that the vertical rings look narrower and longer than the horizontal rings. This illusion is similar to anisotropy for an elongated diamond (Figure 2C & 2D). The diamonds have identical shape but the vertical diamond looks narrower and longer. In Figure 2C & 2D, the angles ϕh and ϕv are physically equal but ϕv looks narrower (the subscript h indicates angles centered around the horizontal axis, while the subscript v indicates angles centered around the vertical axis). Similarly, ψh looks wider despite ψh and ψv being equal angles. As a result, the vertical diamond looks narrower. Cohen and Zaidi (2007) showed that flat 2D angles look broader when the central axis is obliquely oriented compared to vertical and modeled this as an effect of decoding the outputs of V1 cells, where those tuned to horizontal orientations were larger in number and narrower in tuning than those tuned to oblique orientations. We now extend this to comparing vertical versus horizontally centered angles to explain the diamond illusion.

Figure 2. Anisotropy of the shape of the ring.

Figure 2.

A and B correspond to snapshots from Figure 1E and 1F respectively. Despite being physically identical shapes, B is perceived as vertically elongated and narrower in comparison to A. C and D: The shape illusion is similar for two elongated diamonds and can be explained by differences in perceived angles: ϕh is perceived to be wider than ϕv despite being physically equal, and ψh is perceived to be wider than ψv.

Cortical Anisotropy

Figure 3A shows the orientation tuning widths of V1 simple cells we used in our model, centered at the preferred orientation, where 0° and ±180° represent horizontal preferred orientation, and ±90° represent vertical orientation, representing a summary of cat cortex measurements made with drifting gratings (Li et al., 2003), The heights of the curves reflect the relative number of cells tuned to each direction (Li et al., 2003). Cardinal orientations exhibit a higher number of cells and narrower tuning width. However, cells preferring vertical orientation are fewer and broader tuned than cells preferring horizontal orientation. The left two panels in Figure 3B show the Li et al (2003) curves plotted in terms of preferred orientation by averaging across opposite motion directions. There are more recent and more extensive data on primate V1 and V4 (Fang et al., 2022), but we did not use these because they provide only relative numbers, whereas Cohen & Zaidi (2007) found that tuning widths were more important than numbers, and our simulations reproduced that effect. The simulated orientation tuning curves use von Mises probability density functions to generate the tuning curve for all orientations θ:

fθμi,kiexpkicosθ-μi (1)

Figure 3. Cortical anisotropy.

Figure 3.

A: Orientation tuning curves of V1 simple cells were simulated by von Mises distributions to match the anisotropic and isotropic cortices and weighted by the number of cells (See text for references). B: Detailed parameters for the von Mises distributions are shown on the left two panels for the anisotropic (red) and isotropic (black) cortex. Decoded angles around the vertical and horizontal diverging axes respectively are shown on the right two panels. C: The decoded angle difference ( γˆH-γˆV ) from the anisotropic cortex (red) and the isotropic cortex (dotted black). For the isotropic cortex, there is no difference in the decoded angles for γV and γH. However, for the anisotropic cortex, γH is decoded to be broader than γV.

Where μi is the preferred orientation of the i-th cell and ki sets the tuning width. Since we will be decoding angles, two orientations 180° apart are considered distinct, so Equation 1 covers 360°, unlike most fits of the von Mises function (Swindale,1998). As a control, we also decoded angles from a cortex with isotropic distribution of orientation preference and tuning (Figure 3B).

Perceived angle decoded from cortical responses

In Figure 2, angles around the horizontal axis γh are perceived to be wider than equal angles around the vertical axis γv and we decoded the difference between them for angles ranging from 10–170 degrees from an anisotropic cortex as well as an isotropic cortex. We assume that each angle γ is formed by two lines θa and θb, to which the response of ni neurons with preferred orientation μi are rai and rbi:

rai=nifθaμi,kirbi=nifθbμi,ki (2)

Each response is normalized by a divisive gain control:

Rai=raigi,Rbi=rbigiwheregi=rai2+rbi2+c (3)

In primary visual cortex, neuronal responses are suppressed by the presence of surrounding stimuli. This suppression is orientation-specific, with the strongest suppression occurring when the stimuli in the receptive field and the surround share a similar orientation, and the weakest suppression occurring when they are orthogonal to each other (DeAngelis et al., 1994; Slllito et al., 1995; Shushruth et al., 2013). The divisive gain control mechanism, involving self-normalization or dividing by the weighted sum of all other neurons (where the weight is highest for neurons with the same orientation preference), can achieve this neural interaction (Schwartz et al., 2001).

The line orientations, θa and θb, are decoded using vector sums (Georgopoulos et al., 1986; Schwartz et al., 2009):

θa^=AngiRail,θb^=AngiRbil (4)

Where l is the unit vector that points to the preferred orientation. Then, the decoded angle is simply calculated from the two decoded line orientations:

γ^=mod(θa^-θb^,180) (5)

Right two panels in Figure 3B show decoded angles for the anisotropic cortex (red) and the isotropic cortex (black). Figure 3C displays the decoded difference (γh^-γv^) between an angle centered around the horizontal axis and a physically identical angle around the vertical axis, for the anisotropic cortex (red) and the isotropic cortex (black).

Decoded responses from the isotropic cortex show no difference between the estimated horizontal and vertical angles. However, when decoded from the anisotropic cortex, angles around the horizontal axis γh are estimated to be wider than physically equal angles around the vertical axis γv which is consistent with what is observed in Figure 2C & 2D. The difference varies with the magnitude of the angle but is greater than 5 degrees for all but very obtuse angles. We present this simulation to show that the illusory shape anisotropy in diamonds can be attributed to anisotropy in cortical populations. The estimated difference for a physical angle of 90° also explains why a square rotated 45° appears as a diamond which is elongated vertically (Mach, 1897). Similar, but more complex decoding would yield a similar conclusion for the ellipses in Figure 2A & 2B. The qualitative inference is sufficient for our purposes, because the anisotropy assumptions from cat cortex are qualitatively similar but not quantitatively identical to measurements from primate cortex. Without the normalization in Equation 3, the magnitude of the decoded difference is appreciably smaller. Cohen & Zaidi (2007) reproduced perceived differences in vertical versus oblique angles with maximum likelihood decoding by explicitly including cross-orientation suppression.

Effect of shape matching on perceived non-rigidity

Maruya & Zaidi (2024) showed that horizontally rotating elliptical rings with physical aspect ratios greater than 1.0 (narrower than circular rings) are perceived as more non-rigid than horizontally rotating elliptical rings with physical aspect ratios smaller than 1.0 (wider than circular), so the shape difference is a potential explanation for the non-rigidity anisotropy. To test whether this explanation is sufficient, we had observers adjust shapes for vertical and horizontal rotations until they matched, and then tested if matching projected shapes abolishes the non-rigidity anisotropy.

We presented pairs of vertically and horizontally rotating rings and asked observers to match their shapes by stretching one of the pairs horizontally. Note that horizontal stretching makes the horizontally rotating rings narrower, and the vertically rotating rings wider. In Figure 4A, the 1st and 3rd rows show horizontal stretches of the 2D projected image ranging from 0% to 50%, while the 2nd and 4th rows illustrate 3D physical stretches of the rings before projection. Figures 4B and 4C present histograms of the horizontal stretch in the image domain (red) and the physical domain (blue) that match the shapes of horizontally rotating rings (left) and vertically rotating rings (right). To match shapes, observers stretched the horizontal rings and the vertical rings in the horizontal direction by approximately 20–30%. Individual results are shown in Supplementary Figure S1.

Figure 4. Effect of image and physical stretch on perceived non-rigidity.

Figure 4.

A: Examples of rings elongated to match the shape of orthogonally oriented circular rings. Subjects were allowed to stretch rings in image domain (first and third rows) or physically (second and forth rows), ranging from 0% (left column) to 50% (right column). B: Histograms representing the extent of horizontal stretch in the image domain (red) and physical domain (blue) for vertically rotating rings adjusted to match the shape of physically circular horizontally rotating rings. C: Histograms representing the extent of horizontal stretch in the image domain (red) and physical domain (blue) for horizontally rotating rings adjusted to match the shape of physically circular vertically rotating rings. D: Probability of observers reporting vertically rotating rings as more non-rigid. Original: Two pairs of circular rings (black). H indicates that horizontally rotating rings were stretched to match shapes and V indicates that vertically rotating rings were stretched, either in the image red) or physically before projection (blue). Max: Both vertical and horizontal rings were stretched maximally like in the last column of Panel A, so horizontally rotating rings were perceived as narrower and longer than vertically rotating rings.

Next we presented selected pairs of rotating rings in random order and 4 observers reported which pair looked more independent and non-rigid: (i) original circular horizontal versus circular vertical rings, (ii) physically stretched horizontal rings versus circular vertical rings, (iii) physically stretched vertical rings versus circular horizontal rings, (iv) stretched horizontal rings in the image domain versus circular vertical rings, and (v) stretched vertical rings in the image domain versus circular horizontal rings. Figure 4D shows the probability of observers reporting vertically rotating rings as more non-rigid. For the original pair, observers perceived vertically rotating rings as more nonrigid 95.0% of the time (black). Matching the shapes perceptually reduced the nonrigidity anisotropy by up to 40%, with a greater reduction observed in image-stretch (red bars) compared to physical stretch (blue bars). Individual results are shown in Supplementary Figure S2. In an ancillary forced choice experiment, observers picked the image stretch as a better shape match than the physical stretch (61%), possibly because the shape anisotropy is an illusory stretch of the image. However, the non-rigidity anisotropy remains significantly above chance levels despite the shape matches so the perceived shape does not account for almost 60% of the effect. As a critical test, we took the most stretched horizontal and vertical rings (Figure 4A Rightmost rings in 2nd and 4th row) and asked 4 observers to choose which pair was more nonrigid when rotated. Despite the horizontal rings being much narrower than vertical rings, the opposite of the original condition, observers picked the vertical rings as more non-rigid on 80% of the comparisons, opening the possibility that anisotropy of low-level motion detectors also plays a role.

Non-rigidity anisotropy from cortical anisotropy in direction selectivity

As we state in the introduction, since the cat cortical anisotropy was measured with drifting gratings (Li et al., 2003), we make the reasonable assumption that the anisotropic distribution of numbers and tuning widths also represents direction selective cells. To simulate the effect of spatial anisotropy on optic flow, we computed local motion energy for each pixel of the videos. Each motion-energy unit consisted of a quadrature pair of 3D Gabor filters (Watson & Ahumada,1983; Adelson & Bergen, 1985; Watson & Ahumada, 1985; Van Santen & Sperling,1985; Rust, Mante, Simoncelli, & Movshon, 2006; Bradley & Goyal, 2008; Nishimoto & Gallant, 2011; Nishimoto et al., 2011):

Gi,odd(x,y,t)=Qiexp-x-cx,i2+y-cy,i22σs,i2-t-ct,i22σt,i2×sin2πωx0x+2πωy0y+2πωt0t (6)
Gi,even(x,y,t)=Qiexp-x-cx,i2+y-cy,i22σs,i2-t-ct,i22σt,i2×cos2πωx0x+2πωy0y+2πωt0t (6′)

Where Qi is the normalization constant; and cx,i and cy,i are the spatial means of the Gaussian envelopes, and ct,i is the temporal mean; σs,i and σt,i are the spatial and temporal standard deviations of the Gaussian envelopes, respectively; and ωx0, ωy0, and ωt0 are the spatial and temporal frequencies of the sine component of the Gabor filter, referred to as the preferred spatial and temporal frequencies where:

ωx0=F0sinηtcosηs;ωy0=F0sinηtsinηs;ωt0=F0cosηt (7)

ηs and ηt are spatial and temporal orientations (8 spatial orientations from 0 to 7/8π and 5 temporal orientations from -π/4 to π/4) and F0 is the frequency (0.28 and 0.14 cycles/pixel). For the anisotropic cortex, σ is adjusted so that the half-height orientation bandwidths Δω12 are matched to cat cortex estimates (Li et al., 2003) by the following equation (Movellan, 2002):

Δω122tan-10.5F0σ (8)

Each filter Gi,even and Gi,odd was convolved with the video I(x,y,t). The responses of the quadrature pair were squared and summed to produce a phase-independent response mi. Figure 5 A and B illustrate the direction tuning curves for isotropic and anisotropic cortices, weighted by the number of cells. In the anisotropic cortex, cells that prefer a vertically moving bar (either upwards or downwards) exhibit narrower tuning and are more numerous compared to cells that prefer oblique or horizontally moving bars.

Figure 5. Comparing optic flows from isotropic cortex and anisotropic cortex to templates for rotation and wobbling.

Figure 5.

A: Tuning curves for direction selective cells reflecting documented anisotropies in width and number of cells. B: Tuning curves for direction selective cells for an isotropic cortex. C: Optic flow fields generated by isotropic cortex (top row) and anisotropic cortex (bottom row) for physical horizontal rotation (left column) and vertical rotation (right column). Anisotropic cortex generates vectors pointing horizontally. D: Depiction of rotation and wobbling axes. E: Physical velocity field for horizontal (top) and vertical (bottom) rotations mixed with physical wobbling (Weights k=0–1). F & G Cosine similarity of ME optic flows from isotropic (F) and anisotropic (G) cortex to different templates (k=0: physical rotation, k=1: physical wobbling, shown in E). Isotropic cortex leads to no difference between best fitting k for horizontal (blue) and vertical (red) rotation. However, cortical anisotropy results in different best fitting k: more wobbling for vertical rotation.

To obtain the best estimate of velocity (uˆ,vˆ) per point using the whole family of filters, we computed optic flow by minimizing a weighted average of the combined filter responses (Heeger 1988):

u^,v^=argminu,vini(miml¯Niu,vNl¯u,v)2 (9)
Ni(u,v)=exp-4π2σs4σt2uωx0+vωy0+ωt02u2σs2σt2+v2σs2σt2+σs4 (9′)

Where Ni(u,v) is the predicted response of filter i to velocity (u,v), mi is the motion energy output, and ml¯ and Nl¯(u,v) are the sum of Ni and the sum of mi for filters sharing the same spatial orientation. Figure 5C shows optic flow fields for different conditions of rotation of a single ring, horizontal rotation Ry around the y-axis (left column), and vertical rotation Rx around the x-axis (right column), for the isotropic cortex (top row), and the anisotropic cortex (bottom row). The color of the vector corresponds to the direction and the length represents the magnitude of the estimated velocity. Most of the optic flow vectors from the isotropic cortex are orthogonal to the contour. In contrast, for the anisotropic cortex, the optic flow vectors shift towards the horizontal direction, as indicated by the red and turquoise colors.

To understand what this horizontal shift in the velocity field implies for perceived rotation versus wobble (which adds rotation Rz around the z-axis), we derived the velocity field for different combinations of physical rotation and wobbling of a single ring, with the wobbling weight k ranging from k=0~rotation to k=1~wobbling. We began with the vector P(θ)=(cosθ,sinθ,0)T which defines the position of every point on a ring lying on the X-Z plane for each angular coordinate of the circular ring θ (Figure 5D) , and calculated Pk(θ,t) for every combination of rotation and wobble given by Ry and Rz the rotational matrices around the Y and Z axis respectively, with the angular velocity ω and the tilt of the ring, τ:

Pk(θ,t)=Proj(Ry(ωt)Rz(τ)Ry(-kωt)P(θ))=(u(θ,t),v(θ,t))T (10)

Proj(·) is a projection function and we used the orthographic projection. Then, the velocity field Fk is calculated by the partial derivative with respect to time:

Fk=tPk(θ,t) (11)

Figure 5E illustrates the velocity fields for horizontal rotation (top) and vertical (bottom), obtained with wobbling weights (k) increasing from left to right, from pure rotation to pure wobbling. For horizontal rotation (top row), horizontal vectors are predominant when k=0 and vertical vectors at k=1, with the opposite for the vertical rotation (bottom row). These velocity fields serve as templates, and the cosine similarity between these derived templates and the estimated optic flows in Figure 5C is calculated at each point and time for both the isotropic and anisotropic cortex (Figures 5F & 5G) and then averaged over all points for a complete cycle. The blue trace represents mean cosine similarity for horizontal rotation, and the red trace for vertical rotation. As expected, for flows calculated from an isotropic cortex, cosine similarities are the same for horizontal and vertical rotations with the best fit at k=0.84. For flows from the anisotropic cortex, horizontal and vertical rotations result in different cosine similarities. The horizontal rotation cosine similarity curve is relatively flat from k values 0.0 to 0.7 with a peak around 0.4, while the vertical rotation curve is steep and peaks at a higher k, close to 0.9. Thus, if generated velocity fields were compared to velocity field templates for physical rotation and wobbling, it would, explain why observers perceive more non-rigidity when the two rings rotate vertically. However, having all possible templates is biologically implausible, so next we explore how a more biologically plausible factoring of the velocity fields can lead to similar comparisons that predict the non-rigidity anisotropy.

Non-rigidity anisotropy from biologically plausible filters for differential invariants

In kinematics, which describes the motion of objects without considering the forces that cause them to move, motion fields are analyzed in terms of the local differential invariants, divergence, curl, and deformation after excluding translation parallel to the image plane, and these analyses have been useful in understanding object motion from optic flow (Koenderink and van Doorn 1976; Longuet-Higgins and Prazdny 1980; Koenderink, 1986; Todorović, 1993). For the single ring with some combination of rotation and wobble, the divergence field is perpendicular to the contour and captures expansion/contraction of the projection (Figure 6A); the curl field is tangential to the contour and captures rotation; the deformation fields captures shear caused by the difference between orthogonal axes in contraction/expansion or shear and can be combined into one measure giving direction and magnitude of deformation (Koenderink, 1986). Besides providing understandable summary measures, these operators also have biological validity. Zhang et al. (1993) demonstrated that position-independent detectors of curl and divergence can emerge through simple Hebbian learning, from MT-like receptors to MST-like receptors. Physiological studies on nonhuman primates and imaging studies on humans have identified neurons in areas MST, STPa, and STS that preferentially respond to the curl and div components of optic flow (Duffy et al., 1997; Saito et al., 1986; Lappe et al., 1996; Anderson et al., 1999). In addition, neural activity sensitive to def has been found in dorsal MST (Sugihara et al., 2002), and human psychophysical experiments have demonstrated that perceived slant is proportional to def (Koenderink and Doorn., 1976; Freeman et al., 1996; Domini and Caudek, 2003).

Figure 6. Gradients of ME flow from the rotating ring align with percepts.

Figure 6.

A: Examples of differential invariants. B : Gradients of rotation to wobbling, divergence (left), curl (middle), and deformatiosn1 (right) as a function of rotational phase (0°–180°: half a cycle). These gradients illustrate that with increasing emphasis on wobbling from k=0 (blue curves) to k=1 (red curves), variability increases across all three gradients, while rotational fields consistently exhibit higher values of curl and deformation. C: Isotropic Tuning Gradients for Vertical (left) and Horizontal (right) Rotation: Similar gradients are observed for both orientations, with optimal fitting by curves at a wobbling weight of k=0.78. D: Upon adjusting filter numbers and tuning widths to match V1 anisotropies before velocity field calculations, gradients for vertical rotation closely correspond to physical wobbling (k=0.91), while those for horizontal rotation indicate lower levels of wobb ling (k=0.48). E & F : Best k as a function of image stretch. E for isotropic cortex and F for anisotropic cortex. A greater amount of stretch leads to increased wobbling during horizontal rotation (red) and reduced wobbling during vertical rotation (blue). For an anisotropic cortex, the higher amount of wobbling during vertical rotation is maintained, consistent with experimental results. Furthermore, at maximum stretch, the isotropic and anisotropic cortices predict different outcomes: the isotropic cortex suggests that horizontal rotation should be perceived as more non-rigid, while the anisotropic cortex suggests th at vertical rotation should appear more non-rigid.

To derive divergence, curl, and deformation fields for physical rotation and physical wobbling, the dot product of the velocity fields derived in Equation 11 as a function of k (illustrated in Figure 5E) was taken with each operator corresponding to those illustrated in Figure 6A, and then integrated across space to obtain one number for each invariant per phase of motion, where i is the basis vector (1,0) and j is (0,1) :

Divergence:

Fkdvdθi-dudθjdθ (12)

Curl:

Fkdudθi+dvdθjdθ (13)

Deformation:

Def1=Fk-dudθi+dvdθjdθ,Def2=Fkdvdθi+dudθjdθ (14)

These are combined into one measure of the magnitude of Def:

Def=Def12+Def22 (14’)

Figure 6B displays the values of derived integrated gradients for divergence, curl, and deformation from left to right as a function of the motion phase. The colors represent different wobbling weights (k), with warmer colors indicating more wobbling and cooler colors representing more rotation. For all three gradients, the velocity field for rotation exhibits lower variance across different motion phases because the local velocity vectors consistently point in horizontal directions, so changes in the value result solely from alterations in shape due to projection. The wobbling field exhibits higher variance because the projected shapes and thus the local velocities change more over time. The value of Divergence varies from positive to negative representing divergence versus convergence, and the amplitude is greater for wobble because of the additional axis of rotation. Curl for the rotational velocity field is high, but almost constant, reflecting the minimal change in projected shape, whereas the greater projected shape changes for wobbling lead to smaller values of curl but a larger amplitude of variation. The derived deformation estimates have a similar variation. Taken together, the three derived gradients for the physical motions provide templates for distinguishing between relative weights of rotation and wobble, so next we estimated the values of the gradients for the optic flow fields for the single rotating ring computed from the isotropic and anisotropic cortex (Figure 5C). For cortical isotropy (Figure 6C), the divergence, curl, and deformation are identical for vertical and horizontal orientation (red symbols represent divergence, green represent curl, blue represent deformation). The dashed smooth curve in the corresponding color indicates the best fit of the derived physical motion. For both orientations, the best fits occur at a wobbling weight of k=0.78, consistent with the computational results in Maruya and Zaidi (2024 JOV) using CNNs that showed that motion-energy flows support percepts of wobble, because they predominantly signal motion orthogonal to contours. The cortical anisotropy distorts the optic flow so that the optimal fit for vertical rotation (k=0.91) corresponds closer to physical wobbling than the fit (k=0.48) for horizontal rotation. The analysis in terms of differential invariants thus provides an explanation for the anisotropy of non-rigidity that matches the one provided by template matching.

To further test the efficacy of this model, we tested if it could predict reduced anisotropy due to horizontal shape stretching in the shape matching experiment. We took the videos of the rotating single ring and applied horizontal image stretches ranging from 0% to 50%, and calculated the optic flows generated by the isotropic and anisotropic cortex similar to that shown in Figure 5C. Then we calculated the three differential invariants at each image stretch and estimated the optimal k as in Figures 6C and 6D for horizontal rotation and vertical rotation. For both isotropic (Figure 6E) and anisotropic (Figure 6F) cortices, increased stretching during horizontal rotation (red) results in a higher amount of wobbling, while increased stretching during vertical rotation (blue) leads to a reduced amount of wobbling, consistent with the experimental result that reduced anisotropy is observed with a greater amount of horizontal stretch. However, for the anisotropic cortex vertical rotation maintains higher non-rigidity across different amounts of stretch, which aligns with the results in Figure 4D, while the isotropic cortex predicts the opposite. Moreover, the isotropic cortex simulations predict least anisotropy at 0% stretch and most at 50%, which is the opposite of the experimental results, while the anisotropic predicts that the anisotropy reduces with stretch but does not disappear, consistent with the experimental results. Taken together, the simulations from the anisotropic cortex explain the experimental results.

Experimental methods

Stimulus generation

Using Python, we created two circular rings connected rigidly at a 60 degree angle, rotating around either a vertical axis or a horizontal axis oblique to both rings. The joint’s speed when passing fronto-parallel was 6°/s, with a ring diameter of 3 degrees of visual angle (dva), resulting in an angular speed of 0.3 cycles per second (cps). We presented pairs of vertically and horizontally rotating rings and asked observers to match their shapes. The rings could be stretched horizontally from 0% to 50%, either in the image domain or the physical domain. Horizontal stretches in the image domain transformed the projected ellipses for the horizontal rotation into narrower and longer ellipses, and into wider and shorter ellipses for the vertically rotating axes. In the physical domain, stretches transformed the circular rings into elongated ellipses, and transformations in the projected images were similar to those for the image stretches. Figure 4A shows examples of elongated rings as a function of horizontal stretch in the image domain (first and third rows) and the physical domain (second and fourth rows). The matching was done while the rings were rotating in the same phase.

Stimulus display

The videos were displayed on a VIEWPixx/3D (VPixx Technologies, Saint-Bruno-de-Montarville, QC, Canada) at 120 Hz. PsychoPy was used to display the stimulus and run the experiment. The data were analyzed using Python. An observer’s viewing position was fixed with a chin rest to ensure the video was viewed at the same elevation as the camera position.

Psychophysical procedures

Observers were asked to use the mouse to select the best-matched shape. After three trials, the median stretch was taken, and each matched shape in the image and physical domains was presented side by side for the observer to select the most similar shape. Finally, we presented a vertically rotating ring pair and a horizontally rotating ring pair side by side, and observers were asked to select the most non-rigidly moving pair. There were 5 conditions as shown in Figure 4D. The set of conditions was repeated 40 times. Measurements were made by four observers with normal or corrected-to-normal vision. Observers gave written informed consent. All experiments were conducted in compliance with a protocol approved by the institutional review board at SUNY College of Optometry, adhering to the tenets of the Declaration of Helsinki.

Discussion

Despite the abundance of non-rigid 3D objects and organisms in the world, and despite the evolutionary advantage of being able to judge their actions from shape deformations, non-rigid motion perception is an understudied phenomenon (He et al., 2023; Ullman, 1984; Sinha & Poggio,1996; Johansson, 1973;Jain & Zaidi, 2011; Choi et al., 2024; Sparrow et al., 1998; Cohen et al., 2010; Chaung et al., 2012; Braunstein et al., 1990; Vuong et al., 2012). Maruya and Zaidi (JOV 2024) showed that the horizontal rotation of two rigidly connected circular 3D rings can appear either rigid or nonrigid (rings wobbling and rolling independently) if the rotational speed is sufficient to activate motion-energy cells. They found that the nonrigid perception arises from populations of motion energy cells signaling motion orthogonal to the contour. The nonrigidity can be decreased and even abolished if the rings contain salient features that can be used for feature-tracking and if the shape does not support a prior for spinning motion (Maruya & Zaidi, 2024). We strengthen this link by showing that the perceived nonrigidity is strongly affected by the cortical anisotropy for motion-direction selective neurons.

The nonrigidity anisotropy we demonstrate is as interesting as it is unexpected. We show that the hard-wired cortical anisotropy is powerful enough in modifying the optic flow from the vertically rotating rings towards the horizontal directions to counter the effects of salient features that evoke feature-tracking and overcome the effects of shape-based priors in the horizontal rotation (Video in Figure 1H).

A model that explains the complete percept of the 3D rings is beyond the scope of this paper, so we concentrate on the perception of non-rigidity versus rigidity and find surprisingly that variations in this high-level percept can be explained by anisotropies in the neural population, possibly even at the primary visual cortex. That certain variations in a high-level percept such as rigidity are a function of hardwired low-level neural properties is remarkable. Other than Maxwell’s success in identifying cone photoreceptors from color matches, there are almost always too many degrees of freedom to model anything definitive about neural mechanisms from psychophysical results. In addition, it seems that there is little information about higher-level neural processes in lower areas (Xiao et al., 2024). We may have succeeded because we picked a single higher-level aspect of a percept and a well-documented property of visual cortex, so our strategy may be worth repeating.

To link motion-energy to anisotropy we found a combination of computer simulations and mathematical derivations to be extremely useful. For calculating the optic flow for rings rotating horizontally and vertically we used banks of motion-energy filters, and adjusted their spatial envelopes to compare the effects of V1 cortical anisotropy to isotropy. To understand the physical motions that best describe the optic flows, we mathematically derived the velocity fields of rings undergoing different weighted combinations of rotation and wobbling motion (given by the parameter k), and correlated them with the optic flows. For optic flows generated from an isotropic cortex, the best fitted weights were the same for horizontal and vertical rotation. However, the change in the optic flow caused by an anisotropic cortex predicted different physical combinations for horizontal and vertical rotations. Vertical rotation aligned closely with physical wobbling, while horizontal rotation aligned closer to physical rotation, matching the perceptions reported by our observers. We realized that correlating different templates for different motions to the optic flow from an object is computationally expensive and biologically implausible, not to mention that the templates would not generalize across object shapes. Therefore, we explored a more efficient and plausible model using differential invariants.

Differential invariants from kinematics have the advantage of not being restricted to precise shapes, similar to direction-selective filters being more general than matching specific features across frames, and they can be calculated with combinations of biologically plausible filters. Consequently, we derived the differential invariants mathematically for the physical combinations of rotation and wobbling and plotted their values as a function of motion phase. Then we computed the values of the differential invariants given by the optic flows obtained from isotropic and anisotropic cortex and compared them to the values for different combinations of rotation and wobbling. The results strengthen our contention that the anisotropy non-rigidity arises from the hardwired cortical anisotropy in numbers and tuning widths of direction selective neurons.

In summary, our study demonstrates that vertically rotating rings appear more non-rigid than horizontally rotating rings. The shape of the vertically rotating rings seems narrower and longer compared to the horizontally rotating rings. By matching the shapes, observers were able to reduce perceived anisotropy by up to 30%, yet vertically rotating rings were still perceived as more non-rigid after shapes were matched. We incorporated anisotropy into our motion-energy computations and found that the extracted optic flow patterns from anisotropic cortex align with templates for physical wobbling for the vertically rotated rings but not for the horizontally rotated rings. Matching to derived differential invariants as a biologically plausible model, strengthened our conclusions.

Supplementary Material

Supplement 1

Acknowledgement

NEI grants EY035085 & EY035838

Appendix

Stimulus Generation and projection

We created 3D circular, rotating/wobbling rings by applying the equations for rotation along the X,Y,andZ axes in a 3D space to all points on the rendered object. The rotational matrix around each axis is given by:

RX(θ)=1000cosθ-sinθ0sinθcosθ (A1)
Ry(θ)=cosθ0sinθ010-sinθ0cosθ (A2)
RZ(θ)=cosθ-sinθ0sinθcosθ0001 (A3)

If P(θ)=(cosθ,sinθ,0)T is the initial location of a 3-D point on the object lying on the XY plane with radius 1, the position of the point (Pr) on a rotating ring inclined at an angle of τ from the ground plane and angular velocity of ω is expressed as:

Pr=Ry(ωt)Rz(τ)P(θ) (A4)

The position of the point on the wobbling object will be:

Pw=Ry(ωt)RZ(τ)RZ(-ωt)P(θ) (A5)

The operations of the rotational matrix for wobbling and rotation are illustrated in Figure 5 D. In Equation 10, if k=0 in Rz(-kωt), then Rz(-kωt) is the identity matrix. Thus, Equation 10 provides a smooth transition from rotation to wobbling. Then, we projected each point in perspective for the stimulus, but for the analysis, we used orthographic projection. For the physical stretch, P(θ)=(a*cosθ,b*sinθ,0)Ta and b were changed to make changes in aspect ratio that led to narrower or wider ellipses. Image stretches were performed on the projected image.

To compute velocity fields, the time derivative was taken as shown in Equation 11.

Fk=tPk(θ,t)=S(0,1,0)Ry(ωt)Rz(τ)Rz(-kωt)+Ry(ωt)Rz(τ)S(0,-1,0)Rz(-kωt)Pk(θ,t) (A6)

where S(0,1,0) is a skew symmetric matrix 001000-100.

Equation symbol lookup table

Symbol Description
ϕv, ϕh, ψv, ψh Angles in two elongated diamonds depicted in Figure 2C and D
ki, μi Kappa that sets the tuning width and the mean of von Mises function for i-th neuron
γh, γv Angles around the horizontal axis and the vertical axis
θa, θb Orientation of two lines that form an angle of γ
ni Number of cells for i-th neuron
rai, rbi Feedforward response of i-th cell given orientations of θa and θb respectively
gi Divisive gain control of i-th cell
Rai, Rbi Normalized response of i-th cell
θa^, θb^ Decoded orientations
γv^, γh^ Decoded angles given γh and γv respectively
Gi,odd, Gi,even Odd and even 3D Gabor filters of i-th cell
Qi Normalization constant for the Gabor filter
cx,i,cy,i,ct,i Spatial and temporal means of the Gabor filter
σs,i,σt,i Spatial and temporal envelopes of the Gabor filter
ωx0, ωy0, ωt0 Spatial and temporal frequencies of the sine component of the Gabor filter
ηs, ηt Spatial and temporal orientations of the Gabor filter
F0 Frequency of the Gabor filter
Δω12 Half-height orientation bandwidths
uˆ,vˆ The best estimate of local velocity components
Ni(u,v) Predicted response of filter i to velocity (u,v)
mi i-th motion energy output
ml-,Nl-(u,v) The sum of Ni and the sum of mi for filters sharing the same spatial orientation
Ry, Rz Rotational matrices around Y and Z axis respectively
P(θ) The position of a point on the ring lying on the X – Z plane
τ Tilt of the ring
ω Angular velocity of the rotation of the ring
k The wobbling weight (k=0rotation, k=1wobbling)
Proj() The projection function
Pk(θ,t) The position of a point at a space and time for a moving ring with the wobbling weight of k
u(θ,t), v(θ,t) The x and y components for Pk(θ,t)
Fk The velocity field for the moving ring with the wobbling weight of k

Footnotes

The authors declare no competing interests.

References

  1. Adelson E. H., & Bergen J. R. (1985). Spatiotemporal energy models for the perception of motion. Journal of the Optical Society of America A, 2(2), 284–299. [DOI] [PubMed] [Google Scholar]
  2. Aggarwal J. K., Cai Q., Liao W., & Sabata B. (1998). Nonrigid motion analysis: Articulated and elastic motion. Computer Vision and Image Understanding, 70(2), 142–156. [Google Scholar]
  3. Maruya Akihito, Qasim Zaidi; Perceptual transitions between object rigidity and non-rigidity: Competition and cooperation among motion energy, feature tracking, and shape-based priors. Journal of Vision 2024. A; 24(2):3. 10.1167/jov.24.2.3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Maruya Akihito, Zaidi Qasim. (2024B). Anti-Sensory Visual Illusion. I-Perception. [Google Scholar]
  5. Anderson K. C., & Siegel R. M. (1999). Optic flow selectivity in the anterior superior temporal polysensory area, STPa, of the behaving monkey. Journal of Neuroscience, 19(7), 2681–2692. [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Appelle S. (1972). Perception and discrimination as a function of stimulus orientation: The “oblique” effect in man and animals. Psychological Bulletin, 78, 266–278. [DOI] [PubMed] [Google Scholar]
  7. Bigelow A., Kim T., Namima T., Bair W., & Pasupathy A. (2023). Dissociation in neuronal encoding of object versus surface motion in the primate brain. Current Biology, 33(4), 711–719. [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Bradley D. C., & Goyal M. S. (2008). Velocity computation in the primate visual system. Nature Reviews Neuroscience, 9(9), 686–695. [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Braunstein M. L., Hoffman D. D., & Pollick F. E. (1990). Discriminating rigid from nonrigid motion: Minimum points and views. Perception & Psychophysics, 47(3), 205–214. [DOI] [PubMed] [Google Scholar]
  10. Burge T. (2010). Origins of objectivity. Oxford University Press. [Google Scholar]
  11. Choi R., Feldman J., & Singh M. (2024). Perceptual Biases in the Interpretation of Non-Rigid Shape Transformations from Motion. Vision, 8(3), 43. [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Churchland A. K., Gardner J. L., Chou I. H., Priebe N. J., & Lisberger S. G. (2003). Directional anisotropies reveal a functional segregation of visual motion processing for perception and action. Neuron, 37(6), 1001–1011. [DOI] [PubMed] [Google Scholar]
  13. Chuang L. L., Vuong Q. C., & Bülthoff H. H. (2012). Learned non-rigid object motion is a view-invariant cue to recognizing novel objects. Frontiers in Computational Neuroscience, 6, 26. [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Cohen EH, Zaidi Q. Fundamental failures of shape constancy resulting from cortical anisotropy. J Neurosci. 2007. Nov 14; 27(46):12540–5. doi: 10.1523/JNEUROSCI.4496-07.2007. [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Cohen E. H., Jain A., & Zaidi Q. (2010). The utility of shape attributes in deciphering movements of non-rigid objects. Journal of Vision, 10(11), 29–29. [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Coletta N. J., Segu P., & Tiana C. L. (1993). An oblique effect in parafoveal motion perception. Vision Research, 33(18), 2747–2756. [DOI] [PubMed] [Google Scholar]
  17. Dakin S. C., Mareschal I., & Bex P. J. (2005). An oblique effect for local motion: Psychophysics and natural movie statistics. Journal of vision, 5(10), 9–9. [DOI] [PubMed] [Google Scholar]
  18. DeAngelis G. C., Freeman R. D., & Ohzawa I. Z. U. M. I. (1994). Length and width tuning of neurons in the cat’s primary visual cortex. Journal of neurophysiology, 71(1), 347–374. [DOI] [PubMed] [Google Scholar]
  19. Domini F., & Caudek C. (2003). 3-D structure perceived from dynamic information: A new theory. Trends in Cognitive Sciences, 7(10), 444–449. [DOI] [PubMed] [Google Scholar]
  20. Duffy C. J., & Wurtz R. H. (1997). Medial superior temporal area neurons respond to speed patterns in optic flow. Journal of Neuroscience, 17(8), 2839–2851. [DOI] [PMC free article] [PubMed] [Google Scholar]
  21. Dvoeglazova M., & Sawada T. (2024). A role of rectangularity in perceiving a 3D shape of an object. Vision Research, 221, 108433. [DOI] [PubMed] [Google Scholar]
  22. Fang C., Cai X., & Lu H. D. (2022). Orientation anisotropies in macaque visual areas. Proceedings of the National Academy of Sciences, 119(15), e2113407119. [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Freeman T. C., Harris M. G., & Meese T. S. (1996). On the relationship between deformation and perceived surface slant. Vision Research, 36(2), 317–IN1. [DOI] [PubMed] [Google Scholar]
  24. Georgopoulos A. P., Schwartz A. B., & Kettner R. E. (1986). Neuronal population coding of movement direction. Science, 233(4771), 1416–1419. [DOI] [PubMed] [Google Scholar]
  25. Girshick A. R., Landy M. S., & Simoncelli E. P. (2011). Cardinal rules: visual orientation perception reflects knowledge of environmental statistics. Nature Neuroscience, 14(7), 926–932. [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. Griffiths A. F., & Zaidi Q. (2000). Perceptual assumptions and projective distortions in a three-dimensional shape illusion. Perception, 29(2), 171–200. [DOI] [PubMed] [Google Scholar]
  27. Gros B. L., Blake R., & Hiris E. (1998). Anisotropies in visual motion perception: a fresh look. JOSA A, 15(8), 2003–2011. [DOI] [PubMed] [Google Scholar]
  28. Harrison W. J., Bays P. M., & Rideaux R. (2023). Neural tuning instantiates prior expectations in the human visual system. Nature Communications, 14(1), 5320. [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. He D.; Nguyen D.T.; Ogmen H.; Nishina S.; Yazdanbakhsh A. Perception of Rigidity in Three- and Four-Dimensional Spaces. Front. Psychol. 2023, 14, 1180561. [DOI] [PMC free article] [PubMed] [Google Scholar]
  30. Heeger D. J. (1988). Optical flow using spatiotemporal filters. International journal of computer vision, 1, 279–302. [Google Scholar]
  31. Jain A., & Zaidi Q. (2011). Discerning nonrigid 3D shapes from motion cues. Proceedings of the National Academy of Sciences, 108(4), 1663–1668. [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Kendall D, Barden D, Carne T, Le H (1999) Shape and shape theory (Wiley, New York: ). [Google Scholar]
  33. Koenderink J. J. (1986). Optic flow. Vision Research, 26(1), 161–179. [DOI] [PubMed] [Google Scholar]
  34. Koenderink J. J., & van Doorn A. J. (1976). Local structure of movement parallax of the plane. JOSA, 66(7), 717–723. [Google Scholar]
  35. Lappe M., Bremmer F., Pekel M., Thiele A., & Hoffmann K. P. (1996). Optic flow processing in monkey STS: a theoretical and experimental approach. Journal of Neuroscience, 16(19), 6265–6285. [DOI] [PMC free article] [PubMed] [Google Scholar]
  36. Li B, Peterson MR, Freeman RD (2003) Oblique effect: a neural basis in the visual cortex. J Neurophysiol 90:204–217. [DOI] [PubMed] [Google Scholar]
  37. Longuet-Higgins H. C., & Prazdny K. (1980). The interpretation of a moving retinal image. Proceedings of the Royal Society of London. Series B. Biological Sciences, 208(1173), 385–397. [DOI] [PubMed] [Google Scholar]
  38. Lorenceau J., & Shiffrar M. (1992). The influence of terminators on motion integration across space. Vision Research, 32(2), 263–273. [DOI] [PubMed] [Google Scholar]
  39. Lu Z. L., & Sperling G. (2001). Three-systems theory of human visual motion perception: review and update. JOSA A, 18(9), 2331–2370. [DOI] [PubMed] [Google Scholar]
  40. Mach E. (1886). Beiträge zur Analyse der Empfindungen. Jena: Gustav Fischer. In: Williams C. M. (Trans.), English translation: Contributions to the analysis of the sensations (p. 1897). Chicago: The Open Court. [Google Scholar]
  41. Mach E. (1897). Contributions to the Analysis of the Sensations. Open Court. [Google Scholar]
  42. Movellan J. R. (2002). Tutorial on Gabor filters. Open source document, 40, 1–23. [Google Scholar]
  43. Maunsell J. H., & Van Essen D. C. (1983). Functional properties of neurons in middle temporal visual area of the macaque monkey. I. Selectivity for stimulus direction, speed, and orientation. Journal of neurophysiology, 49(5), 1127–1147. [DOI] [PubMed] [Google Scholar]
  44. Nishimoto S., & Gallant J. L. (2011). A three-dimensional spatiotemporal receptive field model explains responses of area MT neurons to naturalistic movies. Journal of Neuroscience, 31(41), 14551–14564. [DOI] [PMC free article] [PubMed] [Google Scholar]
  45. Nishimoto S., Vu A. T., Naselaris T., Benjamini Y., Yu B., & Gallant J. L. (2011). Reconstructing visual experiences from brain activity evoked by natural movies. Current Biology, 21(19), 1641–1646. [DOI] [PMC free article] [PubMed] [Google Scholar]
  46. Oktaviana A. A., Joannes-Boyau R., Hakim B., Burhan B., Sardi R., Adhityatama S., … & Aubert M. (2024). Narrative cave art in Indonesia by 51,200 years ago. Nature, 1–5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  47. Pizlo Z., & De Barros J. A. (2021). The concept of symmetry and the theory of perception. Frontiers in Computational Neuroscience, 15, 681162. [DOI] [PMC free article] [PubMed] [Google Scholar]
  48. Rokem A., & Silver M. A. (2009). A model of encoding and decoding in V1 and MT accounts for motion perception anisotropies in the human visual system. Brain research, 1299, 3–16. [DOI] [PubMed] [Google Scholar]
  49. Rokers B., Yuille A., & Liu Z. (2006). The perceived motion of a stereokinetic stimulus. Vision Research, 46(15), 2375–2387. [DOI] [PubMed] [Google Scholar]
  50. Rust N. C., Mante V., Simoncelli E. P., & Movshon J. A. (2006). How MT cells analyze the motion of visual patterns. Nature Neuroscience, 9(11), 1421–1431. [DOI] [PubMed] [Google Scholar]
  51. Saito H. A., Yukie M., Tanaka K., Hikosaka K., Fukada Y., & Iwai E. (1986). Integration of direction signals of image motion in the superior temporal sulcus of the macaque monkey. Journal of Neuroscience, 6(1), 145–157. [DOI] [PMC free article] [PubMed] [Google Scholar]
  52. Schwartz O., Sejnowski T. J., & Dayan P. (2009). Perceptual organization in the tilt illusion. Journal of Vision, 9(4), 19–19. [DOI] [PMC free article] [PubMed] [Google Scholar]
  53. Schwartz O., & Simoncelli E. P. (2001). Natural signal statistics and sensory gain control. Nature Neuroscience, 4(8), 819–825. [DOI] [PubMed] [Google Scholar]
  54. Shiffrar M., & Pavel M. (1991). Percepts of rigid motion within and across apertures. Journal of Experimental Psychology: Human Perception and Performance, 17(3), 749. [DOI] [PubMed] [Google Scholar]
  55. Shushruth S., Nurminen L., Bijanzadeh M., Ichida J. M., Vanni S., & Angelucci A. (2013). Different orientation tuning of near-and far-surround suppression in macaque primary visual cortex mirrors their tuning in human perception. Journal of Neuroscience, 33(1), 106–119. [DOI] [PMC free article] [PubMed] [Google Scholar]
  56. Sinha P., & Poggio T. (1996). Role of learning in three-dimensional form perception. Nature, 384(6608), 460–463. [DOI] [PubMed] [Google Scholar]
  57. Slllito A. M., Grieve K. L., Jones H. E., Cudeiro J., & Davls J. (1995). Visual cortical mechanisms detecting focal orientation discontinuities. Nature, 378(6556), 492–496. [DOI] [PubMed] [Google Scholar]
  58. Sparrow J. E., & Stine W. W. (1998). The perceived rigidity of rotating eight-vertex geometric forms: extracting nonrigid structure from rigid motion. Vision Research, 38(4), 541–556. [DOI] [PubMed] [Google Scholar]
  59. Dakin Steven C., Mareschal Isabelle, Bex Peter J.; An oblique effect for local motion: Psychophysics and natural movie statistics. Journal of Vision 2005; 5(10):9. 10.1167/5.10.9. [DOI] [PubMed] [Google Scholar]
  60. Sugihara H., Murakami I., Shenoy K. V., Andersen R. A., & Komatsu H. (2002). Response of MSTd neurons to simulated 3D orientation of rotating planes. Journal of Neurophysiology, 87(1), 273–285. [DOI] [PubMed] [Google Scholar]
  61. Todorović D. (1993). Analysis of two-and three-dimensional rigid and nonrigid motions in the stereokinetic effect. JOSA A, 10(5), 804–826. [Google Scholar]
  62. Tomasi C., & Kanade T. (1992). Shape and motion from image streams under orthography: a factorization method. International Journal of Computer Vision, 9, 137–154. [Google Scholar]
  63. Ullman S. (1984). Maximizing rigidity: the incremental recovery of 3-D structure from rigid and nonrigid motion. Perception, 13(3), 255–274. doi: 10.1068/p130255. [DOI] [PubMed] [Google Scholar]
  64. Van Santen J. P., & Sperling G. (1985). Elaborated Reichardt detectors. Journal of the Optical Society of America A, 2(2), 300–321. [DOI] [PubMed] [Google Scholar]
  65. Vezzani S., Kramer P., & Bressan P. (2014). Stereokinetic effect, kinetic depth effect, and structure from motion. In Wagemans J. (Ed.), The Oxford handbook of perceptual organization (pp. 521–540). Oxford, UK: Oxford Library of Psychology. [Google Scholar]
  66. Vuong Q. C., Friedman A., & Read J. C. (2012). The relative weight of shape and non-rigid motion cues in object perception: A model of the parameters underlying dynamic object discrimination. Journal of Vision, 12(3), 16–16. [DOI] [PubMed] [Google Scholar]
  67. Watson A. B., & Ahumada A. (1983). A look at motion in the frequency domain. Silicon Valley, CA: Ames Research Center, National Aeronautics and Space Administration. [Google Scholar]
  68. Watson A. B., & Ahumada A. J. (1985). Model of human visual-motion sensing. Journal of the Optical Society of America A, 2(2), 322–342. [DOI] [PubMed] [Google Scholar]
  69. Weiss Y., & Adelson E. H. (2000). Adventures with gelatinous ellipses—constraints on models of human motion analysis. Perception, 29(5), 543–566. [DOI] [PubMed] [Google Scholar]
  70. Xiao W., Sharma S., Kreiman G., & Livingstone M. S. (2024). Feature-selective responses in macaque visual cortex follow eye movements during natural vision. Nature Neuroscience, 1–10. [DOI] [PMC free article] [PubMed] [Google Scholar]
  71. Zhang K., Sereno M. I., & Sereno M. E. (1993). Emergence of position-independent detectors of sense of rotation and dilation with Hebbian learning: an analysis. Neural Computation, 5(4), 597–612. [Google Scholar]
  72. Zaidi Q., & DeBonet S. J. (2000). Motion energy versus position tracking: Spatial, temporal and chromatic parame [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplement 1

Articles from bioRxiv are provided here courtesy of Cold Spring Harbor Laboratory Preprints

RESOURCES