Abstract
Perceptual gaps can be caused by objects in the foreground temporarily occluding objects in the background or by eyeblinks, which briefly but frequently interrupt visual information. Resolving visual motion across perceptual gaps is particularly challenging, as object position changes during the gap. We examine how visual motion is maintained and updated through externally-driven (occlusion) and internally-driven (eyeblinks) perceptual gaps. Focusing on both phenomenology and potential mechanisms such as suppression, extrapolation, and integration, we present a framework for how perceptual gaps are resolved over space and time. We finish by highlighting critical questions and directions for future work.
Keywords: motion, occlusion, eyeblinks, extrapolation, prediction, suppression
The challenge of processing motion through perceptual gaps
Incoming visual information is often interrupted by externally- and internally-driven events such as occlusion (see Glossary) and eyeblinks, producing perceptual gaps (Figure 1). For example, cars moving on a highway are often fully or partially occluded by other traffic. Similarly, eyeblinks, which can occur 15–20 times per minute [1] and may last for up to 500 ms [2,3], prevent visual information from reaching the brain. One way of bridging these perceptual gaps is to bias perception towards the representation of the last visible information prior to the gap [4]. Retaining information about object identity, for example, can be achieved by relying on the object features last seen [5], as object identity rarely changes during the gap. In contrast, retaining visual motion through the gap is challenging, as a moving object changes position while out of sight. Thus, the visual system needs to update position information during perceptual gaps. To be able to update object position, the direction of motion and the rate of movement needs to be maintained to successfully estimate when and where an object will re-appear. Together, updating position and maintaining a representation of velocity allows for resolving visual motion through perceptual gaps which is critical to prepare for actions, navigate the environment successfully, and avoid collisions.
Recent developments in neuroimaging and behavioural paradigms have increased our ability to investigate the mechanisms for updating motion information during perceptual gaps. For example, advances in analysis methods for time-resolved neuroimaging methods such as electroencephalography (EEG) and magnetoencephalography (MEG) enable a deeper understanding of inherently dynamic processes [6] and have recently been employed to examine processes such as motion extrapolation [7,8]. In addition, modelling techniques have been used to examine how these processes could be instantiated in the brain [9].
Using eyeblinks and occlusion as models, the current review draws on evidence from behavioural and neuroscience studies to discuss how visual motion information is maintained and updated through externally- and internally-driven perceptual gaps. While saccades also interrupt visual information and share some of the characteristics and mechanisms of bridging perceptual gaps caused by eyeblinks and occlusion, we will not focus on this special case in detail (Box 1). We begin by reviewing physical motion and illusory motion processing, which we discuss in the context of perceptual gaps. Then we consider the experience of visual motion during occlusion and eyeblinks and identify possible mechanisms that support motion processing through these gaps. We conclude that motion perception before and after the perceptual gap is critical to make accurate predictions and integrate and resolve perceptual gaps across space and time.
BOX 1: Perceptual gaps caused by saccadic eye-movements.
Disruptions in visual sensory input can also arise from saccadic eye-movements. A saccade is a ballistic eye-movement where fixation is transferred from one spatial location to another, enabling rapid sampling of our visual environment. While both saccades and eyeblinks produce perceptual gaps through physical movement of the eye, we focus on eyeblinks in our review for two reasons. First, eyeblinks cause a physical obstruction between the retina and sensory input with the closing of the eyelid, similar to externally-driven perceptual gaps with object occlusion. In contrast, during a saccade, light still impinges on the retina and visual processing is thought to be inhibited through saccadic suppression. However, sensory input inhibition is incomplete. With training, suppression can actually be dampened if stimulus presentation during a saccade is task relevant [133,134]. In fact, moving task relevant stimuli during saccade flight can be detected and used to support post-saccadic gaze correction [135]. Second, during a saccade, the spatial position of objects needs to be updated for both the saccadic eye-movement and the external motion adding significant complexity to comprehending motion processing during such perceptual gaps. Integration of information across saccades could be supported by receptive field remapping, where neurons with receptive fields which that are about to receive salient information due to a saccade activate start responding in expectation [136,137]. Remapping may also occur for the relocation of attention across saccades, adding complexity for determining which receptive fields are prospectively activated. Information about an attended object may also be maintained across saccade, adding a role for trans-saccadic memory [138]. However, despite these differences, it is notable that eyeblinks and saccades may have similar underlying mechanisms which bridge the perceptual gap they cause. These include suppression around the eye-movement and prediction of post eye-movement sensory input via prior knowledge.
The neural substrates of motion processing
Both physical and illusory motion processing can potentially inform our understanding of how visual motion is resolved through perceptual gaps. Before and after the perceptual gap, the object moves visibly along a motion trajectory. Processing physical motion before the gap informs predictions about the object’s position during the gap. Furthermore, physical motion after the gap can reinforce the perception of an object trajectory retrospectively. Maintenance of a motion trajectory before and after occlusion has such a strong impact that if the object that passes behind the occluder differs in visual features from the one that emerges at the expected position and time, it is perceived as a single object that changed slightly, and not as two separate objects [10–13]. Further, visual illusions of motion such as apparent motion demonstrate that representations of spatiotemporal content can occur in the absence of explicit visual stimulation. Thus, the mechanisms supporting illusory motion processing may also support motion trajectory maintenance during perceptual gaps.
Physical motion processing
To resolve visual motion across perceptual gaps, physical motion information needs to be integrated over space and time (from before to after the perceptual gap). Visual motion processing is supported by a hierarchy of regions and motion selectivity is observed as early as the retina [14,15] (Figure 2A). Direction-selective neurons in primary visual cortex (V1) process motion as a local, one-dimensional feature and the integration of these local signals into coherent motion patterns is thought to occur in V5/MT (in humans often referred to as hMT) [16]. In addition to V5/MT, areas MST [17], V3A [18], V6 [19] and regions in the intraparietal sulcus (IPS) also represent global motion [20,21]. Global motion signals have even been reported in V2 and V3 [22]. These motion-selective areas also integrate motion over time [23,24]. In particular, the lateral intraparietal area (LIP) within non-human primate IPS has been characterized as temporally integrating motion information over long delays in perceptual decision making tasks [e.g., 25].
Motion processing prior to perceptual gaps may play a role in predicting motion trajectories during perceptual gaps. Motion illusions such as the flash-lag effect [26] demonstrate that we predict the current position of a moving object based on its past trajectory. The flash-lag effect occurs when a stationary object is flashed right next to a moving object and participants (falsely) perceive the position of the stationary object to lag behind the moving object. Transcranial magnetic stimulation (TMS) to V5/MT reduces the flash-lag effect [27], indicating that extrapolation relies on activity in V5/MT. Extrapolating motion trajectories inferred from physical motion into perceptual gaps is likely one of the key processes involved in resolving motion information through gaps.
Illusory motion processing
When objects move out of sight during perceptual gaps such as occlusion, we still have a sense that the object persists and keeps moving [10,28]. Due to the lack of physical motion input during occlusion, perceived motion must be internally-generated or maintained. One of the most striking examples of internally-generated motion is apparent motion where we perceive illusory motion in response to discrete visual stimulation in different positions of the visual field [29] (Figure 2B). For instance, when we watch a movie, we readily perceive fluid motion even though the movie consists of a sequence of stationary images. While there are some differences in the conscious experience of motion induced by apparent motion and perceptual gaps, the maintenance of internally-generated motion signals may rely on similar processes in both cases.
Motion perception in apparent motion closely resembles the perception of physical motion, suggesting we actually perceive a stimulus moving along an illusory motion trajectory [30]. The experience of such apparent motion is so strong and spatiotemporally specific that it can interfere with the detection of another object along the illusory motion trajectory [31–33]. Perception of apparent motion involves similar cortical regions as perception of physical motion. Several studies have shown the involvement of V5/MT during apparent motion [34–37]. For example, single-pulse TMS applied to V5/MT reduced the impact of apparent motion on detection of targets along the apparent motion path [37]. In addition to V5/MT, parietal areas have also been implicated in apparent motion perception. LIP in macaques responds strongly to apparent motion when perceived in the preferred direction of the recorded neurons [38]. There is also evidence that V1 is critical for illusory motion awareness [39–41] with increased activity observed along the apparent motion path [42,43]. As there is no true visual stimulation in the visual field locations corresponding to the apparent motion path, the activation in V1 must be driven by feedback connections from higher level visual areas. Although long-range lateral connections within V1 may also support apparent motion processing [44], they are unlikely to explain all activity. In particular, predictive activity related to the apparent motion illusion has been shown to transfer from V1 in one hemisphere to the other across an eye-movement [45]. This indicates a role for feedback as there are limited callosal connections within V1 [46]. Collectively, these studies demonstrate a position-specific representation for illusory apparent motion stimuli that is similar to physical motion, suggesting the same may exist for internally-maintained moving objects in perceptual gaps.
Although similarities exist, there are also important differences between internally-maintained motion perceived in apparent motion and during perceptual gaps such as occlusion. First, the experience of object persistence through periods of occlusion is spontaneous and automatic, occurring as soon as an object moves out of sight. In contrast, for apparent motion to occur, the object needs to re-appear. Second, when an eyeblink or occlusion occurs, the object disappears gradually, and this may be critical for eliciting the internally-maintained motion. In contrast, apparent motion can be elicited with sudden offsets. These differences in phenomenology suggest that a combination of physical and illusory motion processing is central to resolving perceptual gaps.
The phenomenology of perceptual gaps
In natural vision, perceptual gaps are frequent, but these interruptions do not break the perceived continuity of motion trajectories. For eyeblinks in particular, it has been suggested that the perception of continuity is supported by both the suppression of the visual consequences of an eyeblink [3] and an underestimation of the eyeblink duration [47–50]. In eyeblink suppression, efference copy may trigger preparation for the visual consequences of closing the eyelid by lowering visual sensitivity prior to the gap [3]. Blink suppression is thought to reduce the prominence of the gap so visual information before and after the gap is perceived as one continuous visual event. Suppression alone may not be sufficient because the measured neural suppression of light sensitivity is much less than the perceived change in luminance caused by the closure of the eyelid [3,47,49–51]. Blink duration underestimation may also contribute by minimizing the perceived gap. For stationary stimuli, this underestimation may promote continuous and stable perception. However, for moving stimuli, underestimated blink durations could lead to inaccurate position tracking during perceptual gaps. Furthermore, the extent of blink duration underestimation is not modulated by the actual eyeblink duration [47], making position estimation of moving stimuli more challenging. Collectively, while suppression and underestimated blink duration perception likely contribute to minimizing non-informative sensory input, they are not sufficient for resolving visual motion across perceptual gaps because position information needs to be updated.
One way of updating position information during perceptual gaps is to use the physical motion information before the gap to predict the trajectory during the gap. Extrapolated trajectories have been shown to extend into perceptual gaps. For example, position information is represented when an object dynamically moves into the retinal blind spot [52,53]. Similarly, when an eyeblink occurs [54], recent evidence suggests that extrapolation occurs involuntarily into perceptual gaps (Figure 3). Participants observed an object moving smoothly on a trajectory and when they blinked the object would disappear. When reporting the object’s last visible position prior to the eyeblink, participants reported the position to be displaced in the direction of motion, suggesting that motion trajectories were extrapolated. However, when the object just disappeared at a random location in the absence of an eyeblink, there was no overshoot, indicating that the extrapolation is not linked to disappearance of the object per se. Thus, the gradual disappearance of the object may be an important feature when extrapolating trajectories into perceptual gaps. When an object suddenly vanishes, the perceived last visual position is not consistent with the extrapolated motion trajectory [55,56], suggesting that sudden offsets overwrite extrapolated positions [7,57–61]. In contrast, when an object disappears gradually, extrapolation seems to extend into the perceptual gap, as the gradual nature of disappearance makes it more likely that the object continues to exist [28]. Importantly, extrapolation may not bridge the entire perceptual gap. In an eyeblink paradigm [54], participants perceived continuity in motion when the moving object jumped backward during the eyeblink. This might reflect the fact that extrapolation only covers the perceived duration of the gap, which is shorter than the physical duration. Thus, while extrapolation likely enables motion information to be extended into the perceptual gap, it may not be enough to bridge larger gaps.
To be able to update positions during perceptual gaps, velocity information must be maintained. Infants have been shown to be able to consistently incorporate velocity information into the prediction of object-reappearance after occlusion [62,63]. Similarly, the perceived position of moving objects during eyeblinks has been shown to be influenced by velocity, although the estimation of velocity may be inaccurate [54]. It seems that this ability to maintain velocity information is mainly a perceptual feature of the visual system and not a motor feature, as ocular pursuit velocity decreases substantially while tracking an object during occlusion [64]. However, when contrasting groups who are free to engage in ocular pursuit versus groups who fixate at the re-appearance location, there is an advantage in estimating when the object will re-appear for the ocular pursuit group [65]. This suggests that ocular pursuit prior to occlusion does help to maintain velocity information through the gap. However, second order derivatives of velocity (i.e., acceleration) do not seem to be updated during occlusion, even if acceleration before occlusion is psychophysically detectable [66].
Motion trajectories across perceptual gaps are perceived to be valid when the object’s position after the gap is consistent with the physical motion trajectory observed before the gap. However, when an object after occlusion re-appears in an unexpected position or at an unexpected time, object continuity is interrupted [67–72]. This highlights that the predicted motion trajectory is integrated with the physical motion after the gap. As small prediction errors can occur when extrapolating motion, for example, over long distances or when distractors are present [73], small differences between the predicted and actual position after re-appearance should not interrupt the perception of a valid trajectory. In non-human primates, repeated small, systematic differences in time or position result in an adaptation to the latency or endpoint of saccades predicting the re-appearance of the object behind an occluder [74]. Similarly, if an object is systematically displaced during an eyeblink, humans’ post-eyeblink fixation adjusts towards the displaced position within ~35 trials [75,76]. This indicates that the process of motion extrapolation incorporates feedback from the re-appearance after perceptual gaps, highlighting that prediction occurs within the context of integrating physical motion trajectories across the gap.
Internally maintained motion trajectories are not only anchored by physical motion perception before and after the perceptual gap but are also influenced by the physical presence of the occluder. In apparent motion paradigms, it has been shown that large temporal or spatial gaps erode the perception of smooth motion trajectories [77,78]. However, the presence of an occluder can maintain the perception of smooth motion trajectories despite large gaps [79]. In addition, the shape of the occluder has been shown to influence perceived illusory trajectories [80,81]. When a curved occluder is presented adjacent to apparent motion, the illusory moving object is often perceived to follow a curved trajectory [81]. The likelihood of perceiving the curved illusory trajectory increases as time between disappearance and re-appearance of the object increases, suggesting that perception of the illusory trajectory changes, based on the time that has passed since the object disappeared. These findings demonstrate that the physical presence of the occluder plays a key role in maintaining motion trajectories.
In sum, resolving visual motion through perceptual gaps likely involves a combination of suppression, extrapolation, and integration. However, there are differences when comparing mechanisms which contribute to externally- or internally-driven perceptual gaps. Some of these differences relate to the nature of the interruption caused by blinks versus occlusion. For example, the duration of eyeblinks is relatively constant [47], while the duration of occlusion varies with the size of the occluder and the speed of motion. Similarly, the speed of gradual disappearance and re-appearance is skewed for eyeblinks with the gradual disappearance taking a short time and reappearance taking longer (~50 ms to get from open to closed and ~300 ms to get from closed to 97% open) [2]. In contrast, when the speed of object motion remains the same, the gradual disappearance and re-appearance of an object during occlusion happen at the same rate. Finally, we can adjust blink frequency and blink timing in response to task demands [82], allowing for minimal loss of relevant information. In contrast, occluding events can generally not be avoided, as external occluders are rarely under the control of the observer.
Mechanisms supporting motion processing through perceptual gaps
Where is object information maintained during perceptual gaps?
Multiple motion-sensitive cortical regions have been implicated in maintaining representations during perceptual gaps. Similar to representations during apparent motion [83,84], position-specific information has been shown to be represented in early visual cortex during perceptual gaps [85–87]. In one fMRI study [87], participants viewed an object moving on a circular trajectory that was dynamically occluded in one quadrant. Within retinotopically defined areas of V1, V2, and V3 that corresponded to the position of the occluder, the results showed higher activation during occlusion in comparison to a control condition in which the object disappeared (Figure 4). Interestingly, the signal evoked by the occluded object in early visual regions was not shape-specific suggesting only position information was maintained behind the occluder. This is consistent with behavioural findings from the occlusion and eyeblink literature, showing that participants are often unaware of color or shape changes during perceptual gaps [10,12,88,89]. The lack of identity information during occlusion might point to a difference between representations during apparent motion and perceptual gaps, as object position and identity have been shown to be represented in V1 during apparent motion [90]. However, identity information must be retained to some degree, as temporal frequency, for example, has been demonstrated to influence perceived duration during occlusion [91]. One possibility is that identity information of occluded objects is represented in areas other than early visual cortex. For example, a non-human primate study reported that neurons in the banks of the superior temporal sulcus respond when people, but not other objects, become occluded [92].
Human neuroimaging and non-human primate physiology studies also show a contribution of areas along the dorsal visual pathway [93], presumably capturing extrapolated motion paths across the perceptual gap [94–99]. For example, a study on non-human primates examined posterior parietal cortex responses during visual motion tracking with a joystick [96]. The animals had to track a moving object which was briefly occluded in a subset of trials. The results showed that neurons in LIP displayed direction-selective activity for both occluded and visible objects while neurons within MST only responded to the object when it was visible. Relatedly, results of an fMRI study [98] on humans revealed activity in IPS and V5/MT when an object is temporarily occluded versus when it suddenly disappeared.
How is motion information maintained through perceptual gaps?
Extrapolation is an inherent part of motion processing. When motion signals are processed, previous trajectory information is still present throughout the visual system while current position information stimulates the retina [100]. Recent time-series neuroimaging work has demonstrated that there is a neural representation of predicted positions of a moving object ahead of time [7,8,101]. For example, the results of an EEG study [7] have shown that the position of an object can be decoded ahead of time when the object follows a predictable motion sequence, showing that position information is available before the object arrives there. Similar automatic anticipatory motion effects can be observed in motion illusions such as the flash-lag effect. Extrapolation is likely the driving mechanism behind the flash-lag illusion [57] and explicit models have related extrapolation to the effect [9,102,103]. In particular, neural populations in an artificial neural network learn to extrapolate without supervision through spike-timing dependent plasticity [9]. Comparing the extrapolation shift observed in the neural network with the magnitude of the flash-lag effect observed in humans [104], shows that the degree of extrapolation is highly similar (Figure 5A). This indicates that extrapolation occurs spontaneously and automatically when a system is introduced to motion, highlighting its central role in motion perception.
There is also evidence from brain stimulation and patient studies that activity in parietal regions is modulated by position prediction effects [105,106]. Inhibitory TMS to the IPS reduces the impact of motion extrapolation on target detection [106]. These parietal effects may be related to the interruption of temporal processing of visual events [107,108], which may support motion extrapolation. Together, these findings suggest extrapolation extends into perceptual gaps and that parietal areas play a central role in supporting the representation of extrapolated motion trajectories.
As outlined in the phenomenology section, sensory suppression may also facilitate motion processing through perceptual gaps. Physiology research on non-human primates showed that neurons in early visual areas (V1-V4) respond differently when a perceptual gap occurs due to an eyeblink versus a sudden artificial gap (i.e., blank screen) [109,110]. When the gap was caused by an eyeblink, the visual response was suppressed in comparison to when the gap was caused by a blank screen. Similarly, human work has shown that even when retinal illumination is held constant during the eyeblink, activity in mid-level visual, parietal and prefrontal areas are suppressed [111]. In a recent patient study [112], intracranial recordings were obtained from visual areas while patients viewed images of different objects (e.g., faces, houses, tools). When comparing the neural data in trials where a spontaneous blink occurred with trials where an artificial gap (blank screen) was inserted, the re-appearance response in a higher-level visual area was suppressed for eyeblinks (Figure 5B).
Although efference copy could trigger blink suppression, the visual characteristics of an object gradually disappearing may also initiate suppressive effects [113]. This is more consistent with the observation of suppression in occlusion which is unrelated to motor movements. When there is no reliable depth information in a visual scene, intersections of the object contours are the only cues for occlusion [114,115]. The contours created by occlusion are not due to the object’s shape but are accidental contours that only appear because of occlusion. In a monkey physiology study, V4 responded strongly to true object contours while suppressing accidental contours created by occlusion of stationary objects [116]. The suppression of these accidental contours in V4 might be involved in the perception of a whole object behind the occluder instead of separate object fragments. It is likely that such suppression occurs in dynamic occlusion too. Together, the gradual nature of occlusion and eyeblinks seem to support the perception of persistence through perceptual gaps by maintaining extrapolated trajectories and suppressing post-blink sensory onsets.
There are several candidate regions that may support the integration of physical motion information with internally-maintained motion trajectories. Regions within the medial temporal lobe (MTL), such as the hippocampus, have been shown to represent temporospatial predictions. For example, visual space is represented within human entorhinal cortex in grid-type cells [117–119] with a subgroup of these cells firing in a direction-selective fashion prior to an upcoming saccade [120]. In line with these predictive signals in MTL, several fMRI studies have demonstrated that the hippocampus carries predictive signals for learnt sequences [121–125]. For example, participants in one study [121] were trained to respond to a visual cue with an action (right or left button press) which was then systematically associated with a specific visual outcome stimulus. During an fMRI scan, the trained participants completed the full trial sequences (i.e., cue + action = outcome), partial trial sequences (i.e., cue + action), and trials where they only saw the visual outcome stimulus. The results showed that the full sequences were represented in hippocampus while the outcome only stimulus could only be read-out from early visual areas. On trials where the full sequence was correctly decodable in hippocampus, the outcome could be predicted more reliably from V1. This suggests that the sequential, learned order of visual events is represented in hippocampus which feeds information about expectations back to visual cortex. These and other findings have led to the suggestion that completion of gaps between places or other events is one of the central functions of the hippocampus [126,127]. As resolving motion information through perceptual gaps relies on predicting sequences of object positions based on physical motion trajectories, the MTL seems like a potential candidate for integrating physical and internally-maintained motion signals through perceptual gaps.
In addition to regions in the MTL, another sub-cortical structure, the cerebellum, may play a role in resolving visual motion through perceptual gaps. The cerebellum is involved in representing temporal information that is relevant for motor and cognitive function [128,129]. During occlusion the cerebellum has been shown to be involved in time prediction [130–132]. For example, the results of an fMRI occlusion study [130] show that a subregion of the cerebellum was recruited when the velocity of an object changed over time. In this experiment, a moving object became dynamically occluded and re-appeared either too early or too late (velocity judgment task) or shifted to the right or the left (spatial judgement task). The posterior cerebellum became active only during the velocity judgment task with functional connectivity analysis showing a stronger interaction with IPS, the frontal eye fields, and MT. This finding suggests that the cerebellum is involved in predicting visual motion when a change over time is likely. Future work will have to examine whether regions in the cerebellum are recruited to resolve visual motion across perceptual gaps independently of task demands.
Concluding remarks
Resolving visual information through perceptual gaps is a fundamental part of everyday vision, as eyeblinks and occlusion occur frequently. Here, we highlighted why moving stimuli pose a particular challenge to bridging perceptual gaps and showed that mechanisms such as suppression, extrapolation, and integration are important to update visual motion through periods of occlusion and eyeblinks.
Generally, maintaining and updating motion information happens automatically and without conscious effort. However, many studies require participants to explicitly perform a task related to the perceptual gap, which may change the nature of the experience. In future work it will be important to test more naturalistic conditions, which will require a focus on neural activity in the absence of explicit behaviour. The focus on early visual areas in previous neuroimaging studies looking at spatial information during perceptual gaps, is partly driven by the inherent limitations of fMRI. Due to the temporal lag in fMRI, the only way of answering whether information is maintained during occlusion is to focus on retinotopically organized areas. In the future, alternative approaches to examining the nature of the object representations beyond early visual regions could involve techniques with high-temporal resolution such as M/EEG or methods with a finer spatial scale such as ultra-high field and laminar fMRI (see Outstanding Questions). M/EEG would allow the examination of object representations during the perceptual gap, as activations at each timepoint are independent. Laminar fMRI could potentially provide layer-specific activations which might make it possible to distinguish between feedforward and feedback information. Ultimately, integrating neural data with behavioural findings will be critical to elucidate how different mechanisms contribute to the perceptual experience.
Outstanding questions.
How do suppression, extrapolation and integration interact during perceptual gaps? How does the representation change if the object never re-appears? How is information before and after occlusion integrated in the absence of clear extrapolation?
What features beyond position are represented during perceptual gaps? How do task demands change the way we resolve perceptual gaps?
Are there representations during perceptual gaps in areas beyond visual cortex? How do different cortical regions interact to process visual motion during occlusion or eyeblinks?
What role do the MTL and cerebellum play in resolving visual motion through perceptual gaps caused by eyeblinks and occlusion?
Highlights.
Perceptual gaps occur frequently due to eyeblinks and occlusion
Processing information through perceptual gaps is particularly challenging for moving stimuli as dynamic information changes while the object is briefly out of sight
Our perceptual experience suggests that stimulus information such as position is maintained throughout the gap
Suppression, extrapolation, and integration all play a role in representing visual motion through perceptual gaps
Acknowledgements:
LT, GE and CIB are supported by the Intramural Research Program of the NIMH (ZIAMH002909). The authors thank Marianne Duyck and Eli Merriam for providing helpful comments on earlier versions of this manuscript.
Glossary
- Accidental contours
True object contours mark the shape of an object. In contrast accidental contours are created by objects being partially occluded, generating a new contour where the objects intersect.
- Apparent motion
The impression of movement produced by the rapid succession of still objects in different locations.
- Blink suppression
Suppression of the experience of an eyeblink and the sensory input prior and during the eyeblink to support continuous visual perception throughout eyeblinks.
- Efference copy
Neural copy of a motor command to prepare for the perceptual consequences of movement.
- Extrapolation
Representing motion trajectories ahead of time to compensate for processing delays.
- Eyeblinks
Closure of eyelids that occurs largely involuntarily (i.e., not requiring conscious attention) but can also be controlled at will. Spontaneous eyeblinks occur ~15–20 times per minute and last ~100–300ms.
- Flash-lag effect
Visual motion illusion that is thought to be driven by extrapolation. While an object is moving, another object is flashed alongside the moving object briefly. Because of extrapolation, the moving object is (wrongly) perceived to be ahead of the flashed object.
- Illusory motion
Perceiving motion when no physical motion occurs. Examples include visual illusions such as apparent motion.
- IPS
Intraparietal sulcus. Region of the parietal lobe that has been shown to be involved in motion processing.
- MST
Medial superior temporal area. MST is part of the motion complex and is involved in higher-level motion processing.
- Occlusion
Period of object invisibility typically due to movement behind another object.
- Perceptual gap
A period of time where visual processing is interrupted by internally-controlled events (e.g. eyeblinks) or external events (e.g. occlusion).
- Physical motion
Motion signals that are evoked by an object changing position over time. We refer to physical motion when the moving object is visible.
- Receptive field remapping
Activation of receptive field prior to saccade, in preparation for stimulation of the receptive field once saccade has landed.
- Salient
properties of an object which attracts attention above other objects.
- V5/MT
Middle temporal area. MT is the motion complex in the primate brain. When referring to humans, MT is often referred to as hMT.
Footnotes
Publisher's Disclaimer: This is a PDF file of an unedited manuscript that has been accepted for publication. As a service to our customers we are providing this early version of the manuscript. The manuscript will undergo copyediting, typesetting, and review of the resulting proof before it is published in its final form. Please note that during the production process errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain.
References
- 1.Cruz AA et al. (2011) Spontaneous eyeblink activity. The ocular surface 9, 29–41 [DOI] [PubMed] [Google Scholar]
- 2.Kwon K-A et al. (2013) High-speed camera characterization of voluntary eye blinking kinematics. Journal of the Royal Society Interface 10, 20130227. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Volkmann FC et al. (1980) Eyeblinks and visual suppression. Science 207, 900–902 [DOI] [PubMed] [Google Scholar]
- 4.Fischer J and Whitney D (2014) Serial dependence in visual perception. Nature neuroscience 17, 738–743 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Liberman A et al. (2014) Serial dependence of position perception. Journal of Vision 14, 1186–1186 [Google Scholar]
- 6.da Silva FL (2013) EEG and MEG: relevance to neuroscience. Neuron 80, 1112–1128 [DOI] [PubMed] [Google Scholar]
- 7.Blom T et al. (2020) Predictions drive neural representations of visual events ahead of incoming sensory information. Proceedings of the National Academy of Sciences 117, 7510–7515 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Robinson AK et al. (2021) Overlapping neural representations for the position of visible and imagined objects. NBDT 4, 1–28 [Google Scholar]
- 9.Burkitt AN and Hogendoorn H (2021) Predictive visual motion extrapolation emerges spontaneously and without supervision at each layer of a hierarchical neural network with spike-timing-dependent plasticity. Journal of Neuroscience 41, 4428–4438. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Burke L (1952) On the tunnel effect. Quarterly Journal of Experimental Psychology 4, 121–138 [Google Scholar]
- 11.Carey S and Xu F (2001) Infants’ knowledge of objects: Beyond object files and object tracking. Cognition 80, 179–213 [DOI] [PubMed] [Google Scholar]
- 12.Flombaum JI et al. (2004) Dynamic object individuation in rhesus macaques: A study of the tunnel effect. Psychological Science 15, 795–800 [DOI] [PubMed] [Google Scholar]
- 13.Michotte A et al. (1964) Les complements amodaux des structures perceptives, Institut de psychologie de l’Université de Louvain. [Google Scholar]
- 14.Nassi JJ et al. (2006) The parvocellular LGN provides a robust disynaptic input to the visual motion area MT. Neuron 50, 319–327 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Sincich LC et al. (2004) Bypassing V1: a direct geniculate input to area MT. Nature neuroscience 7, 1123–1128 [DOI] [PubMed] [Google Scholar]
- 16.Zeki S (2015) Area V5—a microcosm of the visual brain. Frontiers in integrative neuroscience 9, 21. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Helfrich RF et al. (2013) Processing of coherent visual motion in topographically organized visual areas in human cerebral cortex. Brain topography 26, 247–263 [DOI] [PubMed] [Google Scholar]
- 18.Nakhla N et al. (2021) Neural selectivity for visual motion in macaque area V3A. Eneuro 8, doi: 10.1523/ENEURO.0383-20.2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Pitzalis S et al. (2015) The human cortical areas V6 and V6A. Visual neuroscience 32, E007. [DOI] [PubMed] [Google Scholar]
- 20.Konen CS and Kastner S (2008) Representation of eye movements and stimulus motion in topographically organized areas of human posterior parietal cortex. Journal of Neuroscience 28, 8361–8375 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Pitzalis S et al. (2020) Neural bases of self-and object-motion in a naturalistic vision. Human brain mapping 41, 1084–1111 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Furlan M and Smith AT (2016) Global motion processing in human visual cortical areas V2 and V3. Journal of Neuroscience 36, 7314–7324 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Pack CC and Born RT (2001) Temporal dynamics of a neural solution to the aperture problem in visual area MT of macaque brain. Nature 409, 1040–1042 [DOI] [PubMed] [Google Scholar]
- 24.Bair W and Movshon JA (2004) Adaptive temporal integration of motion in direction-selective neurons in macaque visual cortex. Journal of Neuroscience 24, 7305–7323 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Huk AC and Shadlen MN (2005) Neural activity in macaque parietal cortex reflects temporal integration of visual motion signals during perceptual decision making. Journal of Neuroscience 25, 10420–10436 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Nijhawan R (1994) Motion extrapolation in catching. Nature 370, 256–257. [DOI] [PubMed] [Google Scholar]
- 27.Maus GW et al. (2013) The perceived position of moving objects: transcranial magnetic stimulation of area MT+ reduces the flash-lag effect. Cerebral cortex 23, 241–247 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Gibson JJ et al. (1969) The change from visible to invisible. Perception & Psychophysics 5, 113–116 [Google Scholar]
- 29.Wertheimer M (1912) Experimentelle studien uber das sehen von bewegung. Zeitschrift fur psychologie 61, [Google Scholar]
- 30.Kolers PA and von Grünau M (1976) Shape and color in apparent motion. Vision research 16, 329–335 [DOI] [PubMed] [Google Scholar]
- 31.Hidaka S et al. (2011) Inhibition of target detection in apparent motion trajectory. Journal of Vision 11, 2–2 [DOI] [PubMed] [Google Scholar]
- 32.Hogendoorn H et al. (2008) Interpolation and extrapolation on the path of apparent motion. Vision research 48, 872–881 [DOI] [PubMed] [Google Scholar]
- 33.Yantis S and Nakama T (1998) Visual interactions in the path of apparent motion. Nature neuroscience 1, 508–512 [DOI] [PubMed] [Google Scholar]
- 34.Goebel R et al. (1998) The constructive nature of vision: direct evidence from functional magnetic resonance imaging studies of apparent motion and motion imagery. European Journal of Neuroscience 10, 1563–1573 [DOI] [PubMed] [Google Scholar]
- 35.Mikami A (1991) Direction selective neurons respond to short-range and long-range apparent motion stimuli in macaque visual area MT. International Journal of Neuroscience 61, 101–112 [DOI] [PubMed] [Google Scholar]
- 36.Muckli L et al. (2002) Apparent motion: event-related functional magnetic resonance imaging of perceptual switches and states. Journal of Neuroscience 22, RC219–RC219 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Vetter P et al. (2015) TMS over V5 disrupts motion prediction. Cerebral cortex 25, 1052–1059 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Williams ZM et al. (2003) Parietal activity and the perceived direction of ambiguous apparent motion. Nature neuroscience 6, 616–623 [DOI] [PubMed] [Google Scholar]
- 39.Hurme M et al. (2019) V1 activity during feedforward and early feedback processing is necessary for both conscious and unconscious motion perception. NeuroImage 185, 313–321 [DOI] [PubMed] [Google Scholar]
- 40.Pedersini CA et al. (2020) Neural bases of visual processing of moving and stationary stimuli presented to the blind hemifield of hemianopic patients. Neuropsychologia 141, 107430. [DOI] [PubMed] [Google Scholar]
- 41.Silvanto J et al. (2005) Striate cortex (V1) activity gates awareness of motion. Nature neuroscience 8, 143–144 [DOI] [PubMed] [Google Scholar]
- 42.Muckli L et al. (2005) Primary Visual Cortex Activity along the Apparent-Motion Trace Reflects Illusory Perception. PLoS Biology 3, e265. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Sterzer P et al. (2006) Primary visual cortex activation on the path of apparent motion is mediated by feedback from hMT+/V5. Neuroimage 32, 1308–1316 [DOI] [PubMed] [Google Scholar]
- 44.Chemla S et al. (2019) Suppressive traveling waves shape representations of illusory motion in primary visual cortex of awake primate. Journal of Neuroscience 39, 4282–4298 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Edwards G et al. (2017) Predictive feedback to V1 dynamically updates with sensory input. Scientific reports 7, 1–12 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Saenz M and Fine I (2010) Topographic organization of V1 projections through the corpus callosum in humans. Neuroimage 52, 1224–1229 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47.Duyck M et al. (2021) Visual continuity during blinks and alterations in time perception. Journal of Experimental Psychology: Human Perception and Performance 47, 1–12 [DOI] [PubMed] [Google Scholar]
- 48.Duyck M et al. (2015) Does time stop when we blink? Journal of Vision 15, 370–370 [Google Scholar]
- 49.Grossman S et al. (2019) Where does time go when you blink? Psychological science 30, 907–916 [DOI] [PubMed] [Google Scholar]
- 50.Irwin DE and Robinson MM (2016) Perceiving a continuous visual world across voluntary eye blinks. Journal of experimental psychology: human perception and performance 42, 1490. [DOI] [PubMed] [Google Scholar]
- 51.Volkmann FC (1986) Human visual suppression. Vision research 26, 1401–1416 [DOI] [PubMed] [Google Scholar]
- 52.Maus GW and Nijhawan R (2008) Motion Extrapolation Into the Blind Spot. Psychol Sci 19, 1087–1091 [DOI] [PubMed] [Google Scholar]
- 53.Shi Z and Nijhawan R (2012) Motion extrapolation in the central fovea. PloS one 7, e33651. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 54.Maus GW et al. (2020) Perceiving locations of moving objects across eye blinks. Psychological Science. 10.1177/0956797620931365 [DOI] [PubMed] [Google Scholar]
- 55.Kerzel D (2000) Eye movements and visible persistence explain the mislocalization of the final position of a moving target. Vision Research 40, 3703–3715 [DOI] [PubMed] [Google Scholar]
- 56.Whitney D et al. (2000) Illusory spatial offset of a flash relative to a moving stimulus is caused by differential latencies for moving and flashed stimuli. Vision research 40, 137–149 [DOI] [PubMed] [Google Scholar]
- 57.Hogendoorn H (2020) Motion Extrapolation in Visual Processing: Lessons from 25 Years of Flash-Lag Debate. Journal of Neuroscience 40, 5698–5705 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 58.Nijhawan R (2008) Visual prediction: Psychophysics and neurophysiology of compensation for time delays. Behavioral and Brain Sciences 31, 179. [DOI] [PubMed] [Google Scholar]
- 59.Nijhawan R (2002) Neural delays, visual motion and the flash-lag effect. Trends in cognitive sciences 6, 387–393 [DOI] [PubMed] [Google Scholar]
- 60.Maus GW and Nijhawan R (2009) Going, going, gone: Localizing abrupt offsets of moving objects. Journal of Experimental Psychology: Human Perception and Performance 35, 611. [DOI] [PubMed] [Google Scholar]
- 61.Maus GW and Nijhawan R (2006) Forward displacements of fading objects in motion: The role of transient signals in perceiving position. Vision research 46, 4375–4381 [DOI] [PubMed] [Google Scholar]
- 62.Rosander K and von Hofsten C (2004) Infants’ emerging ability to represent occluded object motion. Cognition 91, 1–22 [DOI] [PubMed] [Google Scholar]
- 63.Von Hofsten C et al. (2007) Predictive tracking over occlusions by 4-month-old infants. Developmental Science 10, 625–640 [DOI] [PubMed] [Google Scholar]
- 64.Bennett SJ and Barnes GR (2003) Human ocular pursuit during the transient disappearance of a visual target. Journal of Neurophysiology 90, 2504–2520 [DOI] [PubMed] [Google Scholar]
- 65.Bennett SJ et al. (2010) Eye movements influence estimation of time-to-contact in prediction motion. Experimental brain research 206, 399–407 [DOI] [PubMed] [Google Scholar]
- 66.Benguigui N et al. (2003) Time-to-contact estimation of accelerated stimuli is based on first-order information. Journal of Experimental Psychology: Human Perception and Performance 29, 1083. [DOI] [PubMed] [Google Scholar]
- 67.Falck A et al. (2020) Core cognition in adult vision: A surprising discrepancy between the principles of object continuity and solidity. Journal of Experimental Psychology: General 149, 2250–2263. [DOI] [PubMed] [Google Scholar]
- 68.Flombaum JI et al. (2009) Spatiotemporal priority as a fundamental principle of object persistence. The origins of object knowledge [Google Scholar]
- 69.Flombaum JI and Scholl BJ (2006) A temporal same-object advantage in the tunnel effect: facilitated change detection for persisting objects. Journal of Experimental Psychology: Human Perception and Performance 32, 840. [DOI] [PubMed] [Google Scholar]
- 70.Scholl BJ (2007) Object persistence in philosophy and psychology. Mind & Language 22, 563–591 [Google Scholar]
- 71.Scholl BJ (2001) Spatiotemporal priority and object identity. Cahiers de Psychologie Cognitive, 20(5), 359–371 [Google Scholar]
- 72.Scholl BJ and Pylyshyn ZW (1999) Tracking Multiple Items Through Occlusion: Clues to Visual Objecthood. Cognitive Psychology 38, 259–290 [DOI] [PubMed] [Google Scholar]
- 73.Lyon DR and Waag WL (1995) Time course of visual extrapolation accuracy. Acta psychologica 89, 239–260 [DOI] [PubMed] [Google Scholar]
- 74.Itoh TD et al. (2020) Spatial and temporal adaptation of predictive saccades based on motion inference. Scientific reports 10, 1–11 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 75.Lau WK and Maus G (2018) Anisotropic gaze adaptation in reflexive and voluntary blinks. Journal of Vision 18, 600–600 [Google Scholar]
- 76.Maus GW et al. (2017) Target displacements during eye blinks trigger automatic recalibration of gaze direction. Current biology 27, 445–450 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 77.Burr DC (1979) Acuity for apparent vernier offset. Vision Research 19, 835–837 [DOI] [PubMed] [Google Scholar]
- 78.Gepshtein S and Kubovy M (2007) The lawful perception of apparent motion. Journal of Vision 7, 9–9 [DOI] [PubMed] [Google Scholar]
- 79.Scherzer TR and Ekroll V (2012) Occlusion improves the interpolation of sampled motion. Vision research 62, 17–25 [DOI] [PubMed] [Google Scholar]
- 80.Gheorghes TN et al. (2020) Local biases drive, but do not determine, the perception of illusory trajectories. Scientific reports 10, 1–12 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 81.Kim S-H et al. (2012) Curved apparent motion induced by amodal completion. Attention, Perception, & Psychophysics 74, 350–364 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 82.Hoppe D et al. (2018) Humans quickly learn to blink strategically in response to environmental task demands. Proceedings of the National Academy of Sciences 115, 2246–2251 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 83.Alink A et al. (2010) Stimulus predictability reduces responses in primary visual cortex. Journal of Neuroscience 30, 2960–2966 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 84.Schwiedrzik CM et al. (2007) A spatio-temporal interaction on the apparent motion trace. Vision research 47, 3424–3433 [DOI] [PubMed] [Google Scholar]
- 85.Ban H et al. (2013) Topographic representation of an occluded object and the effects of spatiotemporal context in human early visual areas. Journal of Neuroscience 33, 16992–17007 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 86.de Haas B and Schwarzkopf DS (2018) Spatially selective responses to Kanizsa and occlusion stimuli in human visual cortex. Scientific reports 8, 1–11 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 87.Erlikhman G and Caplovitz GP (2017) Decoding information about dynamically occluded objects in visual cortex. NeuroImage 146, 778–788 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 88.O’Regan KJ et al. (2000) Picture changes during blinks: Looking without seeing and seeing without looking. Visual Cognition 7, 191–211 [Google Scholar]
- 89.O’Regan KJ et al. (1999) Change-blindness as a result of ‘mudsplashes.’ Nature 398, 34–34 [DOI] [PubMed] [Google Scholar]
- 90.Chong E et al. (2016) Reconstructing representations of dynamic visual objects in early visual cortex. Proceedings of the National Academy of Sciences 113, 1453–1458 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 91.Maarseveen J et al. (2017) Representing dynamic stimulus information during occlusion. Vision research 138, 40–49 [DOI] [PubMed] [Google Scholar]
- 92.Baker C et al. (2001) Neuronal representation of disappearing and hidden objects in temporal cortex of the macaque. Experimental Brain Research 140, 375–381 [DOI] [PubMed] [Google Scholar]
- 93.Kravitz DJ et al. (2011) A new neural framework for visuospatial processing. Nature Reviews Neuroscience 12, 217–230 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 94.Assad JA and Maunsell JH (1995) Neuronal correlates of inferred motion in primate posterior parietal cortex. Nature 373, 518–521 [DOI] [PubMed] [Google Scholar]
- 95.Delle Monache S et al. (2017) Differential contributions to the interception of occluded ballistic trajectories by the temporoparietal junction, area hMT/V5+, and the intraparietal cortex. Journal of Neurophysiology 118, 1809–1823 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 96.Eskandar EN and Assad JA (1999) Dissociation of visual, motor and predictive signals in parietal cortex during visual guidance. Nature neuroscience 2, 88–93 [DOI] [PubMed] [Google Scholar]
- 97.Makin AD et al. (2009) Tracking visible and occluded targets: Changes in event related potentials during motion extrapolation. Neuropsychologia 47, 1128–1137 [DOI] [PubMed] [Google Scholar]
- 98.Olson IR et al. (2004) Neuronal representation of occluded objects in the human brain. Neuropsychologia 42, 95–104 [DOI] [PubMed] [Google Scholar]
- 99.Shuwairi SM et al. (2007) Neural substrates of dynamic object occlusion. Journal of cognitive neuroscience 19, 1275–1285 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 100.Kwon O-S et al. (2015) Unifying account of visual motion and position perception. Proceedings of the National Academy of Sciences 112, 8142–8147 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 101.Aitken F et al. (2020) Prior Expectations of Motion Direction Modulate Early Sensory Processing. J. Neurosci 40, 6389–6397 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 102.Khoei MA et al. (2017) The flash-lag effect as a motion-based predictive shift. PLoS computational biology 13, e1005068. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 103.Hogendoorn H and Burkitt AN (2019) Predictive coding with neural transmission delays: A real-time temporal alignment hypothesis. Eneuro 6, DOI: 10.1523/ENEURO.0412-18.2019 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 104.Wojtach WT et al. (2008) An empirical explanation of the flash-lag effect. Proceedings of the National Academy of Sciences 105, 16338–16343 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 105.de Vito S et al. (2015) When brain damage “improves” perception: neglect patients can localize motion-shifted probes better than controls. Journal of neurophysiology 114, 3351–3358 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 106.Edwards G et al. (2017) Predictive position computations mediated by parietal areas: TMS evidence. NeuroImage 153, 49–57 [DOI] [PubMed] [Google Scholar]
- 107.Battelli L et al. (2007) The ‘when’pathway of the right parietal lobe. Trends in cognitive sciences 11, 204–210 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 108.Battelli L et al. (2008) The ‘when’parietal pathway explored by lesion studies. Current opinion in neurobiology 18, 120–126 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 109.Gawne TJ and Martin JM (2000) Activity of primate V1 cortical neurons during blinks. Journal of neurophysiology 84, 2691–2694 [DOI] [PubMed] [Google Scholar]
- 110.Gawne TJ and Martin JM (2002) Responses of primate visual cortical neurons to stimuli presented by flash, saccade, blink, and external darkening. Journal of neurophysiology 88, 2178–2186 [DOI] [PubMed] [Google Scholar]
- 111.Bristow D et al. (2005) Blinking suppresses the neural response to unchanging retinal stimulation. Current Biology 15, 1296–1300 [DOI] [PubMed] [Google Scholar]
- 112.Golan T et al. (2016) Human intracranial recordings link suppressed transients rather than’filling-in’to perceptual continuity across blinks. Elife 5, e17243. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 113.Idrees S et al. (2020) Perceptual saccadic suppression starts in the retina. Nature communications 11, 1–19 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 114.Cavanagh P (1987) Reconstructing the third dimension: Interactions between color, texture, motion, binocular disparity, and shape. Computer Vision, Graphics, and Image Processing 37, 171–195 [Google Scholar]
- 115.Nakayama K et al. (1995) Visual surface representation: A critical link between lower-level and higher-level vision. Visual cognition: An invitation to cognitive science 2, 1–70 [Google Scholar]
- 116.Bushnell BN et al. (2011) Partial occlusion modulates contour-based shape encoding in primate area V4. Journal of Neuroscience 31, 4012–4024 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 117.Killian NJ et al. (2012) A map of visual space in the primate entorhinal cortex. Nature 491, 761–764 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 118.Nau M et al. (2018) How the brain’s navigation system shapes our visual experience. Trends in cognitive sciences 22, 810–825 [DOI] [PubMed] [Google Scholar]
- 119.Nau M et al. (2018) Hexadirectional coding of visual space in human entorhinal cortex. Nature neuroscience 21, 188–190 [DOI] [PubMed] [Google Scholar]
- 120.Killian NJ et al. (2015) Saccade direction encoding in the primate entorhinal cortex during visual exploration. Proceedings of the National Academy of Sciences 112, 15743–15748 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 121.Hindy NC et al. (2016) Linking pattern completion in the hippocampus to predictive coding in visual cortex. Nature neuroscience 19, 665–667 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 122.Kok P et al. (2020) Content-based dissociation of hippocampal involvement in prediction. Journal of Cognitive Neuroscience 32, 527–545 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 123.Kok P and Turk-Browne NB (2018) Associative prediction of visual shape in the hippocampus. Journal of Neuroscience 38, 6888–6899 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 124.Schapiro AC et al. (2012) Shaping of object representations in the human medial temporal lobe based on temporal regularities. Current biology 22, 1622–1627 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 125.Wang R et al. (2017) Learning predictive statistics: strategies and brain mechanisms. Journal of Neuroscience 37, 8412–8427 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 126.Buzsáki G and Tingley D (2018) Space and time: The hippocampus as a sequence generator. Trends in cognitive sciences 22, 853–869 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 127.Friston K and Buzsáki G (2016) The functional anatomy of time: what and when in the brain. Trends in cognitive sciences 20, 500–511 [DOI] [PubMed] [Google Scholar]
- 128.Bareš M et al. (2019) Consensus paper: decoding the contributions of the cerebellum as a time machine. From neurons to clinical applications. The Cerebellum 18, 266–286 [DOI] [PubMed] [Google Scholar]
- 129.Ivry RB and Keele SW (1989) Timing functions of the cerebellum. Journal of cognitive neuroscience 1, 136–152 [DOI] [PubMed] [Google Scholar]
- 130.O’reilly JX et al. (2008) The cerebellum predicts the timing of perceptual events. Journal of Neuroscience 28, 2252–2260 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 131.Cerminara NL et al. (2009) An internal model of a moving visual target in the lateral cerebellum. The Journal of physiology 587, 429–442 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 132.Roth MJ et al. (2013) The cerebellum optimizes perceptual predictions about external sensory events. Current Biology 23, 930–935 [DOI] [PubMed] [Google Scholar]
- 133.Scholes C et al. (2021) Learning to silence saccadic suppression. Proceedings of the National Academy of Sciences 118, [DOI] [PMC free article] [PubMed] [Google Scholar]
- 134.Zimmermann E (2020) Saccade suppression depends on context. Elife 9, e49700. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 135.Schweitzer R and Rolfs M (2021) Intrasaccadic motion streaks jump-start gaze correction. Science Advances 7, eabf2218. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 136.Merriam EP and Colby CL (2005) Active vision in parietal and extrastriate cortex. The Neuroscientist 11, 484–493 [DOI] [PubMed] [Google Scholar]
- 137.Neupane S et al. (2020) Perisaccadic remapping: What? How? Why? Reviews in the Neurosciences 31, 505–520 [DOI] [PubMed] [Google Scholar]
- 138.Edwards G et al. (2018) Decoding trans-saccadic memory. Journal of Neuroscience 38, 1114–1123 [DOI] [PMC free article] [PubMed] [Google Scholar]