Skip to main content
Frontiers in Psychology logoLink to Frontiers in Psychology
. 2013 Jan 21;3:620. doi: 10.3389/fpsyg.2012.00620

Predictive Feedback and Conscious Visual Experience

Matthew F Panichello 1,*, Olivia S Cheung 1, Moshe Bar 1,2
PMCID: PMC3549576  PMID: 23346068

Abstract

The human brain continuously generates predictions about the environment based on learned regularities in the world. These predictions actively and efficiently facilitate the interpretation of incoming sensory information. We review evidence that, as a result of this facilitation, predictions directly influence conscious experience. Specifically, we propose that predictions enable rapid generation of conscious percepts and bias the contents of awareness in situations of uncertainty. The possible neural mechanisms underlying this facilitation are discussed.

Keywords: predictions, context, object recognition, priming, visual awareness, top-down, perception, associative processing

Introduction

Perception has evolved to transform raw sensory signals into information that can guide behavior. But perception is not a purely stimulus-driven phenomenon; the brain perceives the world proactively. Endogenous influences, such as attention, memory, and mood, guide perception to ensure that informative representations of the environment are generated as efficiently as possible.

In this review, we focus on the influence of predictions on visual perception. The predictions we discuss are not analogous to the deliberate, elaborative foresight that individuals may engage in when planning for the future (e.g., thinking about which route home from work will have the least traffic). Instead, we use the term “prediction” here to refer to expectations about the immediate sensory environment based on previous experience and learning. Through everyday experience, individuals learn many types of regularities in the world, such as associations among items or events. For example, a yellow traffic light is almost always followed by a red light. Insofar as the environment is generally regular, it is, to an extent, predictable. Expectations about the environment can be derived from these learned regularities and used to guide sensory processing, presumably via top-down (descending) projections linking brain regions involved in generating expectations with lower-level sensory regions.

If predictions guide processing in visual areas, it follows that this facilitation may have consequences for conscious perception1. We argue that predictive mechanisms may shape the contents of visual awareness during instances of sensory ambiguity, allowing subjective experience of the world to remain informative and coherent. Additionally, when sensory input is less ambiguous, predictions may allow percepts to be generated more quickly and with less interference by sensory noise. In other words, the expectation elicited by a yellow traffic light may cause a driver to consciously perceive the subsequent red light more quickly than otherwise possible. We conclude by discussing the neural mechanisms underlying these predictive influences.

Predictions Influence Conscious Perception

Predictions guide the interpretation of ambiguous stimuli

One method for studying the influence of predictions on awareness is to show observers stimuli with multiple perceptual interpretations and determine if predictions affect what they see. The popular perceptual phenomenon known as binocular rivalry provides a convenient means to accomplish this. Binocular rivalry occurs when a unique image is presented to each eye; interpretation of the stimuli is inherently ambiguous because the two eyes provide conflicting information about the shared portion of the visual field. The brain must work out the most likely interpretation of this impossible input (Hohwy et al., 2008). The resolution of this problem is striking: the two stimuli do not fuse together into a blended percept, but rather alternately dominate perception (for reviews of binocular rivalry and multistable perception more generally, see Leopold and Logothetis, 1999; Blake and Logothetis, 2002; Sterzer et al., 2009).

Binocular rivalry is sensitive to many factors, including low-level stimulus properties (e.g., Kaplan and Metlay, 1964; Fahle, 1982) and higher-order endogenous influences such as attention, imagery, and affect (e.g., Ooi and He, 1999; Pearson et al., 2008; Anderson et al., 2011), so a single process or mechanism should not be credited as the “source” of rivalry (Blake and Logothetis, 2002). However, recent findings suggest that predictions play a role in resolving the ambiguity inherent in this phenomenon. For example, a stimulus will tend to dominate rivalry if it has been presented more frequently than the competing stimulus in the recent past; the likelihood of each stimulus is estimated based on recent experience, generating expectations that can guide perception (Chopin and Mamassian, 2012). The tendency of stimuli to change position smoothly during motion can also provide predictive cues for perception. Thus, observers who view a stream of images depicting a rotating grating are more likely to perceive the stimulus consistent with the rotation trajectory at the onset of rivalry (Figure 1A; Denison et al., 2011). Verbal stimuli can also produce an expectation for semantically related stimuli. To demonstrate this, Costello et al. (2009) used a variant of binocular rivalry, termed continuous flash suppression, which relies on the tendency of high-contrast dynamic noise to dominate awareness when presented in rivalry with another stimulus (Tsuchiya and Koch, 2005). Costello et al. found that a word presented to one eye breaks suppression by high frequency noise presented to the other eye more quickly when a semantically related prime is displayed prior to rivalry (Figure 1B).

Figure 1.

Figure 1

Predictions affect the outcome of binocular rivalry. (A) When a stream of images depicting a grating rotating in 45° increments is presented to both eyes, observers are more likely to perceive the grating consistent with this pattern (the “matching eye” stimulus) at the onset of rivalry. Adapted, under creative commons license, from Denison et al. (2011). (B) A word prime (“salt”) causes a semantically related target word (“pepper”; left eye) to break suppression by dynamic noise (right eye) during rivalry more quickly than when the prime and target are not related. The increasing contrast of the target over time helps ensure that the target eventually breaks suppression. Adapted, with permission, from Costello et al. (2009).

In addition to binocular rivalry, bistable figures reveal the ability of top-down predictions to influence conscious perception. Bistable figures have two mutually exclusive interpretations that alternately dominate awareness during viewing. However, a predictive cue related to one of the interpretations of a bistable figure can bias perception in favor of that interpretation (Figure 2A; Bugelski and Alampay, 1961; Balcetis and Dale, 2007; Goolkasian and Woodberry, 2010). Stimuli can also display bistable motion; an array of moving dot stimuli can be arranged so that observers perceive a rotating cylinder with spontaneous reversals in rotation direction. However, when observers are led to expect that the cylinder will consistently rotate in a particular direction, this interpretation dominates perception (Sterzer et al., 2008).

Figure 2.

Figure 2

Predictions affect the interpretation of ambiguous stimuli. (A) Example of an ambiguous figure with its two possible perceptual interpretations emphasized (left). Viewing a prime related to one of the interpretations (right) biases perception in favor of that interpretation. Adapted, with permission, from Goolkasian and Woodberry (2010). Copyright ©by Springer Publishing. (B) These two disks appear to be concave and convex, respectively, due to learned assumption that light comes from above. Experience violating this assumption causes a new shading scheme to maximally evoke the illusion of contour. Adapted from Adams et al. (2004). Copyright ©by Nature Publishing Group.

Intriguingly, predictions can alter the perception of stimuli that are normally perceptually stable. For example, most individuals see the left disk in Figure 2B as a convex and the right disk as concave. However, the contour of each disk is actually ambiguous: this subjective interpretation is driven by the shading of the figures and the learned assumption that “light comes from above” (Brewster, 1826; Sun and Perona, 1998; Adams et al., 2004). New experience contradicting this assumption can cause a substantial shift in inferred light position; as a result, a new shading scheme will maximally evoke the illusion of contour (Adams et al., 2004). Similarly, individuals learn over time that fast-moving objects are less common than slow moving and stationary objects, producing an expectation that can guide perception. New visual experience dominated by fast-moving stimuli alters this expectation, and by extension, subsequent motion perception (Stocker and Simoncelli, 2006; Sotiropoulos et al., 2011). Together, these studies demonstrate that predictions derived from experience can affect how observers see the world.

Predictions facilitate perception during object recognition

Even when visual input is not ambiguous, predictive mechanisms may influence conscious perception. We hypothesize that predictions allow conscious percepts to be generated more efficiently and with less inhibition by sensory noise2. The most direct evidence for this claim is provided by studies showing that predictions can facilitate perception of objects when visual input is noisy. Observers viewing fragmented object figures are more likely to perceive the objects when they are informed about their category (e.g., “an animal” for an elephant; Reynolds, 1985). Similarly, observers instructed to indicate their subjective visual experience report successfully perceiving expected stimuli at greater levels of degradation or lower contrast than unexpected stimuli (Eger et al., 2007; Esterman and Yantis, 2010; Melloni et al., 2011). In the latter study, electrophysiological activity evoked by expected stimuli correlated with subjective experience earlier than that evoked by unexpected stimuli, suggesting that predictions reduced the latency of neural activity related to conscious perception (Melloni et al., 2011).

A large body of research demonstrates that predictions make object recognition faster and more accurate. These studies provide additional, although indirect, evidence for predictive facilitation of conscious perception. These results must be interpreted with caution for two reasons. First, object recognition is a multifaceted process, with multiple stages that may be amenable to influence by predictive mechanisms (e.g., the assignment of semantic attributes to the perceived object). Second, most of the studies cited below examined the influence of predictions on behavioral proxies of recognition such as reaction time rather than subjective reports of perception. However, it will be worthwhile for future research to determine if predictive facilitation of object recognition is due, at least in part, to faster and more accurate generation of object percepts.

For instance, contextual facilitation of object recognition arises from the knowledge that certain objects often reliably co-occur in particular settings. Cubicles and copy machines, but not octopuses and cars, are found inside office buildings. Furthermore, objects within a context are often arranged in a regular manner: computer monitors rest on desks inside of cubicles. Over time, observers learn many sets of such regularities, termed schemata or “context frames” (Friedman, 1979). Cues within a scene can activate associated context frames, allowing observers to predict other features of the environment (Bar, 2004). Accordingly, observers identify objects faster and more accurately when they are shown in their typical environment (e.g., a toaster in the kitchen, Biederman, 1972; Palmer, 1975; Biederman et al., 1982; Davenport and Potter, 2004) or are preceded by an object drawn from the same context (e.g., a bedroom dresser and a vanity mirror, Gronau et al., 2008; see also Sachs et al., 2011). Conversely, recognition is impaired when the expected spatial relationships among objects in a scene are disrupted (Biederman, 1972).

Associations among stimuli can be reinforced over a lifetime of experience. However, they can also be generated quite quickly under artificial conditions in the laboratory (e.g., Chun and Jiang, 1998, 1999; Aminoff et al., 2007; Kim et al., 2009; den Ouden et al., 2010; Turk-Browne et al., 2010). Thus, associative learning appears to be quite flexible and continuously updated. For instance, in a study of perceptual prediction (den Ouden et al., 2010), observers were asked to judge whether a degraded image depicted a face or a house. An auditory cue presented at the start of each trial signaled which stimulus type was more likely to appear, and the experimenters manipulated the predictive strength of these cues over time. Modeling of the behavioral data demonstrated that observers updated their estimates of stimulus category likelihood on a trial-by-trial basis and responded more quickly when cues were highly predictive.

Apart from learned associations among stimuli, preliminary processing of visual information can also create expectancies that facilitate perception. Visual stimuli contain information distributed across a range of spatial frequencies (Figure 3A). Low spatial frequency (LSF) information is rapidly extracted from incoming sensory input and encodes gross properties such as the global shape of the environment and its constituent objects (Figure 3B). In contrast, high spatial frequency (HSF) information is processed more slowly and corresponds to edges and fine details (Figure 3C; Shapley, 1990; Schyns and Oliva, 1994; Bar et al., 2006). It is possible that LSF information can elicit predictions (Bar, 2003) because objects in the same basic-level category often display a similar global shape (Rosch et al., 1976). For example, individual dogs vary tremendously, but all dogs share roughly the same gross features (e.g., four legs and a tail), which immediately differentiate them from members of many other basic-level categories. With experience, we come to learn the defining features of many kinds of objects. Thus, while LSF information lacks fine detail, it is sufficient to trigger general category information stored in memory that guides interpretation of the stimulus (Bar, 2003; Bar et al., 2006; Oliva and Torralba, 2007). In addition to individual objects, different exemplars of basic-level scenes (such as a city street) also share global features that can constrain and facilitate recognition (Bar, 2004; Oliva and Torralba, 2007). Thus, rapid processing of LSF information may explain the remarkable ability of observers to extract the “gist” of a scene at a glance (Biederman et al., 1974; Thorpe et al., 1996; Oliva and Torralba, 2007).

Figure 3.

Figure 3

An image filtered to include (A) both low and high spatial frequency information (B) predominantly low spatial frequencies (C) predominantly high spatial frequencies. The low spatial frequency information in (B) is not sufficient to identify the polar bear but nevertheless substantially constrains the possible interpretations of the stimulus.

To summarize, predictions may shape the contents of awareness when visual input is ambiguous and enable faster and more sensitive conscious perception under less strenuous circumstances. The ability of predictions to shape perception under conditions of uncertainty is clearly advantageous; it is better to generate a meaningful interpretation of the world, informed by previous experience, than to faithfully represent a noisy sensory signal. When visual input is more informative, predictions still guide sensory processing to make conscious perception as efficient as possible. Note that although we suggest that predictions influence conscious experience, predictions themselves may be learned and applied without awareness (e.g., Chun and Jiang, 1998, 1999; Kim et al., 2009; Turk-Browne et al., 2010). We next consider how, at the neural level, predictions may guide perception.

The Neural Basis of Prediction

Descending cortico–cortico projections allow predictions to guide sensory processing

An increasingly popular framework (e.g., Mumford, 1992; Rao and Ballard, 1999; Friston, 2005; Friston and Kiebel, 2009), termed “predictive coding”, posits that top-down predictions facilitate perception by reducing the need to reconstruct the environment via exhaustive bottom-up analysis of incoming sensory information. Validated predictions efficiently explain away some of the sensory input, conserving resources for analysis of unpredicted components. The hierarchical organization of sensory cortex suggests a natural computational architecture for the integration of top-down predictions and bottom-up sensory information. Ascending projections in sensory cortex are sparse and focused, terminating in predominately layer 4 of cortical targets, whereas descending connections project to a larger number of region and innervate their targets in superficial and deep layers of cortex (Felleman and Van Essen, 1991; Friston, 2009). Predictive coding proposes that descending projections convey predictions about the content or organization of the sensory input. Ascending projections convey any information incongruent with the predictions (i.e., an “error term”) received from higher sensory areas. Thus, reciprocally connected cortical areas are able to engage in a dynamic process in which predictions are modified based on incoming sensory input until the higher-level region is able to arrive at reasonable approximation of the incoming input (Ullman, 1995; Friston, 2005).

This predictive coding framework has traditionally been presented with the hierarchical organization of sensory cortex in mind (e.g., Mumford, 1992; Friston, 2005). However, the existence of reciprocal connections between prefrontal and visual cortices in the macaque (Webster et al., 1994; Cavada et al., 2000) and evidence of functional interaction between prefrontal and inferior temporal regions in humans (e.g., Bar et al., 2006; Kveraga et al., 2007; Axmacher et al., 2008) are well documented. These findings suggest that prefrontal regions provide an additional source of feedback to the visual hierarchy (but note that prefrontal regions do not seem to be arranged hierarchically with respect to each other; Yeo et al., 2011). Accordingly, expectation-based prefrontal modulation of sensory processing has been reported (Bar et al., 2006; Summerfield et al., 2006; Eger et al., 2007; Kveraga et al., 2007; Gamond et al., 2011). Predictive mechanisms may thus be instantiated in a variety of brain regions, but consistently seem to depend on the use of descending connections to allow the dynamic comparison of predictions with sensory input.

Converging imaging and computational evidence supports the proposed role of predictive feedback during perception. If descending predictions efficiently facilitate the interpretation of sensory input, then predicted stimuli should evoke less activity in sensory cortex, consistent with the notion that stimulus-driven activity conveys an “error term” communicating the remainder of the signal that has not been explained by top-down predictions. This effect should be especially pronounced in visual processing regions that are particularly devoted to the processing of the predicted stimuli. Accordingly, shapes that appear in an expected location in the visual field elicit less activity in the retinotopic region V1 (Alink et al., 2010), and expected face and place stimuli elicit less activity in face and place-sensitive regions of high-level visual cortex (Egner et al., 2010).

Moreover, because predictive feedback from higher-level processing regions shapes the activity of lower-level sensory regions, activity in lower-level sensory regions should track surprisingly sophisticated aspects of the sensory input. A subset of retinotopic neurons in visual cortex seem sensitive to stimuli outside of their receptive fields, a property replicated in a computational model of visual cortex allowing for predictive feedback to these cells (Rao and Ballard, 1999). Similarly, activity in low-level retinotopic visual areas varies with image recognition (Hsieh et al., 2010; see also Gorlin et al., 2012), conscious experience of figure-ground segregation (Lamme et al., 1998, 2002), and target detection (Supèr et al., 2001), implicating top-down modulation by higher-order areas.

To summarize, expectations derived from learned regularities are able to guide perception via descending feedback targeting sensory regions. We next elaborate on these ideas by reviewing the neural mechanisms for two previously discussed predictive processes: predictions based on contextual associations and predictions based on preliminary LSF information extracted from stimuli. Different neural generators may support these mechanisms, but evidence suggests that both may require interaction between higher cortical areas and visual cortex to allow top-down predictions to shape sensory processing.

Predictions based on contextual associations among objects

In the previous section, we have suggested that contextual associations among objects can create expectancies that guide perception. A distributed network including parahippocampal cortex (PHC), retrosplenial complex (RSC), and medial prefrontal cortex (mPFC) mediates these contextual associations (Figure 4; Bar and Aminoff, 2003; Kveraga et al., 2011). To localize this context network, activity elicited by objects that are strongly associated with a particular context is compared with that elicited by objects that are only weakly associated with any particular context. For example, golf carts are usually found in the context of a golf course and are thus closely associated with other objects that share this context such as golf clubs and golf balls, while cameras lack strong associations because they are found in a variety of contexts and thus do not consistently appear with any particular set of objects (Bar and Aminoff, 2003; stimuli available at http://barlab.mgh.harvard.edu/ContextLocalizer.htm). Thus, associative processing seems to specifically engage these context network regions. Further support for this claim is provided by the fact that these regions are recruited in a variety of tasks that call on contextual associations, such as memory encoding (Peters et al., 2009), navigation (Rauchs et al., 2008; Brown et al., 2010), and future thought (Szpunar et al., 2009).

Figure 4.

Figure 4

The context network. A contrast of strong and weak context object stimuli, revealing the context network on the left hemisphere of an inflated cortical surface. RSC, retrosplenial complex; PHC, parahippocampal cortex; MPFC, medial prefrontal cortex. Reprinted from Kveraga et al. (2011). Copyright ©by National Academy of Sciences/Highwire Press.

There are several mechanisms by which contextual processing may modulate activity in regions related to visual perception. PHC itself is situated in the ventral visual stream and lesions of this region in the monkey severely impair object recognition (Buckley and Gaffan, 1998; Murray and Mishkin, 1998). Thus, one possibility is that other context network regions facilitate the emergence of situation-specific representations in PHC in a top-down manner (Bar, 2007; Bar et al., 2008). MEG recordings have shown that strong context objects elicit phase-locking (a measure of functional interaction) between PHC and other context network regions as early as 170 ms after stimulus onset (Kveraga et al., 2011), suggesting that the interaction between these regions can occur early enough to influence perception. Alternatively, context network regions may also influence processing in lateral occipital cortex (LOC), a region implicated in object recognition (e.g., Malach et al., 1995; Grill-Spector et al., 2001). Contextual information modulates LOC response during recognition (Altmann et al., 2004; Gronau et al., 2008; MacEvoy and Epstein, 2011), suggesting that this region may be receiving feedback signals from contextual processing regions. Further research will need to clarify the precise neural mechanism by which contextual predictions guide processing in sensory cortex.

Predictions based on low spatial frequency information

The prefrontal cortex, specifically orbitofrontal cortex (OFC), may play a critical role in generating predictions based on LSF information (Bar, 2003; Bar et al., 2006). In this proposed framework, LSF information is extracted from visual input and projected to OFC via the magnocellular cells of the dorsal visual stream, which preferentially respond to and rapidly conduct LSF information (Maunsell et al., 1990; Shapley, 1990; Merigan and Maunsell, 1993; Bullier and Nowak, 1995; Chen et al., 2007). As we have suggested, the coarse representations conveyed by LSFs are sufficient to activate a subset of possible candidates regarding the identity of the visual input (Bar, 2003). These predictions are then projected back to object recognition regions in inferior temporal cortex, facilitating perception (Bar et al., 2006).

Consistent with this proposal, LSF images evoke activity in OFC prior to inferior temporal areas; thus, LSF information reaches prefrontal cortex quickly enough to influence recognition processes in the ventral visual stream (Bar et al., 2006). Furthermore, LSF images elicit significant functional coupling between early visual areas and OFC and between OFC and ventral stream areas (peaking approximately 85 and 135 ms post-stimulus, respectively) while HSF images do not (Bar et al., 2006). These findings are consistent with the rapid transmission of LSF information to OFC, followed by top-down feedback from OFC to the ventral stream.

Facilitatory feedback originating in OFC has also been shown using stimuli designed to preferentially stimulate the dorsal magnocellular pathway (Kveraga et al., 2007). Magnocellular cells are sensitive to small differences in luminance contrast but are insensitive to color, whereas the parvocellular cells known to dominate the ventral stream are sensitive to color but relatively insensitive to luminance (Livingstone and Hubel, 1988). Accordingly, line drawings of objects in which figure and ground are identical in color but slightly different in luminance preferentially stimulate magnocellular cells (i.e., “M-biased” stimuli; Steinman and Steinman, 1997; Cheng et al., 2004; Kveraga et al., 2007). Such stimuli can be used to examine whether the magnocellular pathway is indeed important for conveying information to OFC, enabling top-down facilitation. Indeed, M-biased stimuli preferentially stimulate OFC and elicit functional interaction between early visual areas, OFC, and inferior temporal cortex (Kveraga et al., 2007). Furthermore, the degree of OFC activation elicited by M-biased stimuli is inversely correlated with reaction time for object recognition, suggesting that the processing occurring in OFC is indeed facilitating perception (Kveraga et al., 2007). Intriguingly, despite this facilitation, M-biased stimuli elicited less activity in ventral visual regions than stimuli designed to stimulate parvocellular cells, providing indirect evidence that predictive feedback may reduce the need for exhaustive bottom-up processing during recognition.

We have suggested that LSF representations in OFC trigger associations with object and category information stored in memory, which then serve as predictions that guide sensory processing. To activate memory representations, OFC should interact with hippocampal regions in the medial temporal lobe, known to support long-term memory (for a review, see Squire et al., 2004). Indeed, OFC and medial temporal regions have been shown to be reciprocally connected in non-human primates (Rempel-Clower and Barbas, 2000). In humans, OFC engages in functional coupling with the medial temporal lobe during memory retrieval (Nyberg et al., 1995; Piefke et al., 2003; Tsukiura and Cabeza, 2008; Anderson et al., 2010; Colgin, 2011). It will be important to further clarify the possibility that OFC engages in association-based memory retrieval during visual perception.

Summary and Conclusion

We have reviewed evidence that predictions have consequences for conscious perception. When visual input is ambiguous, predictions may help select the contents of awareness, maintaining a coherent interpretation of the environment. Under less demanding conditions, predictions may still influence awareness, allowing percepts to be generated more quickly and with less interference by sensory noise. To support these arguments, we have drawn on observations from a wide variety of domains, including the resolution of binocular rivalry, perception of ambiguous figures, associative learning, and other phenomena. We suggest, however, that all of these studies index the fact that humans are highly adept at extracting consistencies in the world and using this knowledge to generate expectations about the immediate sensory environment. Although different instances of prediction may recruit different cortical regions, predictive mechanisms are likely instantiated as dynamic top-down modulation of sensory cortex by higher sensory and prefrontal areas engaged in comparatively abstract processing. Via this modulation, predictions about the environment generated in higher-level cortical regions can guide perception.

Given that the specific neural processes that give rise to conscious perception remain unclear, it is difficult to conjecture precisely how predictive feedback influences the contents of awareness. However, it is intriguing that a number of prominent theories posit that top-down feedback may play an important role in generating the neural states postulated to account for consciousness (Tononi and Edelman, 1998; Lamme, 2010; Dehaene and Changeux, 2011; Meyer, 2012). Indeed, disrupting top-down processes seems to impair awareness (Pascual-Leone and Walsh, 2001; Ro et al., 2003; Fahrenfort et al., 2007; Dux et al., 2010). Perhaps predictions play not only a modulatory but a driving role in awareness, particularly when other top-down processes such as attention are not engaged. Future research should explore whether an individual’s threshold for visual awareness increases when predictive processes are impaired, such as in depression in which associative processing may be limited (Bar, 2009).

Conflict of Interest Statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Acknowledgments

This work was supported by NSF grant BCS-0842947 and DARPA grant N10AP20036.

Footnotes

1We use “subjective visual experience,” “conscious perception,” and “visual awareness” interchangeably to refer to the phenomenological experience of seeing, distinct from other forms of consciousness such as self-awareness.

2When interpreting the influence of predictions on perception, it is important to consider potential contributions of attention because the two may be easily confounded. However, attention and prediction are dissociable in several ways (Hohwy, 2012; Summerfield and Egner, 2009) and efforts have begun to elucidate the independent contributions of each to perception (Wyart et al., 2012).

References

  1. Adams W. J., Graf E. W., Ernst M. O. (2004). Experience can change the ‘light-from-above’ prior. Nat. Neurosci. 7, 1057–1058 10.1038/nn1312 [DOI] [PubMed] [Google Scholar]
  2. Alink A., Schwiedrzik C. M., Kohler A., Singer W., Muckli L. (2010). Stimulus predictability reduces responses in primary visual cortex. J. Neurosci. 30, 2960–2966 10.1523/JNEUROSCI.3730-10.2010 [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Altmann C. F., Deubelius A., Kourtzi Z. (2004). Shape saliency modulates contextual processing in the human lateral occipital complex. J. Cogn. Neurosci. 16, 794–804 10.1162/089892904970825 [DOI] [PubMed] [Google Scholar]
  4. Aminoff E., Gronau N., Bar M. (2007). The parahippocampal cortex mediates spatial and nonspatial associations. Cereb. Cortex 17, 1493–1503 10.1093/cercor/bhl078 [DOI] [PubMed] [Google Scholar]
  5. Anderson E., Siegel E. H., Barrett L. F. (2011). What you feel influences what you see: the role of affective feelings in resolving binocular rivalry. J. Exp. Soc. Psychol. 47, 856–860 10.1016/j.jesp.2011.02.009 [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Anderson K. L., Rajagovindan R., Ghacibeh G. A., Meador K. J., Ding M. (2010). Theta oscillations mediate interaction between prefrontal cortex and medial temporal lobe in human memory. Cereb. Cortex 20, 1604–1612 10.1093/cercor/bhp223 [DOI] [PubMed] [Google Scholar]
  7. Axmacher N., Schmitz D. P., Wagner T., Elger C. E., Fell J. (2008). Interactions between medial temporal lobe, prefrontal cortex, and inferior temporal regions during visual working memory: a combined intracranial EEG and functional magnetic resonance imaging study. J. Neurosci. 28, 7304–7312 10.1523/JNEUROSCI.1778-08.2008 [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Balcetis E., Dale R. (2007). Conceptual set as a top – down constraint on visual object identification. Perception 36, 581–595 10.1068/p5678 [DOI] [PubMed] [Google Scholar]
  9. Bar M. (2003). A cortical mechanism for triggering top-down facilitation in visual object recognition. J. Cogn. Neurosci. 15, 600–609 10.1162/089892903321662976 [DOI] [PubMed] [Google Scholar]
  10. Bar M. (2004). Visual objects in context. Nat. Rev. Neurosci. 5, 617–629 10.1038/nrn1476 [DOI] [PubMed] [Google Scholar]
  11. Bar M. (2007). The proactive brain: using analogies and associations to generate predictions. Trends Cogn. Sci. (Regul. Ed.) 11, 280–289 10.1016/j.tics.2007.05.005 [DOI] [PubMed] [Google Scholar]
  12. Bar M. (2009). A cognitive neuroscience hypothesis of mood and depression. Trends Cogn. Sci. (Regul. Ed.) 13, 456–463 10.1016/j.tics.2009.08.009 [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Bar M., Aminoff E. (2003). Cortical analysis of visual context. Neuron 38, 347–358 10.1016/S0896-6273(03)00167-3 [DOI] [PubMed] [Google Scholar]
  14. Bar M., Aminoff E., Schacter D. L. (2008). Scenes unseen: the parahippocampal cortex intrinsically subserves contextual associations, not scenes or places per se. J. Neurosci. 28, 8539–8544 10.1523/JNEUROSCI.0987-08.2008 [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Bar M., Kassam K. S., Ghuman A. S., Boshyan J., Schmid A. M., Dale A. M., et al. (2006). Top-down facilitation of visual recognition. Proc. Natl. Acad. Sci. U.S.A. 103, 449–454 10.1073/pnas.0507062103 [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Biederman I. (1972). Perceiving real-world scenes. Science 177, 77–80 10.1126/science.177.4043.77 [DOI] [PubMed] [Google Scholar]
  17. Biederman I., Mezzanotte R. J., Rabinowitz J. C. (1982). Scene perception: detecting and judging objects undergoing relational violations. Cogn. Psychol. 14, 143–177 10.1016/0010-0285(82)90007-X [DOI] [PubMed] [Google Scholar]
  18. Biederman I., Rabinowitz J. C., Glass A. L., Stacy E. W. (1974). On the information extracted from a glance at a scene. J. Exp. Psychol. 103, 597–600 10.1037/h0037158 [DOI] [PubMed] [Google Scholar]
  19. Blake R., Logothetis N. K. (2002). Visual competition. Nat. Rev. Neurosci. 3, 13–21 10.1038/nrn701 [DOI] [PubMed] [Google Scholar]
  20. Brewster D. (1826). On the optical illusion of the conversion of cameos into intaglios, and of intaglios into cameos, with an account of other analogous phenomena. Edinb. J. Sci. 4, 99–108 [Google Scholar]
  21. Brown T. I., Ross R. S., Keller J. B., Hasselmo M. E., Stern C. E. (2010). Which way was I going? Contextual retrieval supports the disambiguation of well learned overlapping navigational routes. J. Neurosci. 30, 7414–7422 10.1523/JNEUROSCI.2167-10.2010 [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Buckley M. J., Gaffan D. (1998). Perirhinal cortex ablation impairs visual object identification. J. Neurosci. 18, 2268–2275 [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Bugelski B. R., Alampay D. A. (1961). The role of frequency in developing perceptual sets. Can. J. Psychol. 15, 205–211 10.1037/h0083443 [DOI] [PubMed] [Google Scholar]
  24. Bullier J., Nowak L. G. (1995). Parallel versus serial processing: new vistas on the distributed organization of the visual system. Curr. Opin. Neurobiol. 5, 497–503 10.1016/0959-4388(95)80011-5 [DOI] [PubMed] [Google Scholar]
  25. Cavada C., Compañy T., Tejedor J., Cruz-Rizzolo R. J., Reinoso-Suárez F. (2000). The anatomical connections of the macaque monkey orbitofrontal cortex. A review. Cereb. Cortex 10, 220–242 10.1093/cercor/10.3.220 [DOI] [PubMed] [Google Scholar]
  26. Chen C., Lakatos P., Shah A. S., Mehta A. D., Givre S. J., Javitt D. C., et al. (2007). Functional anatomy and interaction of fast and slow visual pathways in macaque monkeys. Cereb. Cortex 17, 1561–1569 10.1093/cercor/bhl067 [DOI] [PubMed] [Google Scholar]
  27. Cheng A., Eysel U. T., Vidyasagar T. R. (2004). The role of the magnocellular pathway in serial deployment of visual attention. Eur. J. Neurosci. 20, 2188–2192 10.1111/j.1460-9568.2004.03675.x [DOI] [PubMed] [Google Scholar]
  28. Chopin A., Mamassian P. (2012). Predictive properties of visual adaptation. Curr. Biol. 22, 622–626 10.1016/j.cub.2012.02.021 [DOI] [PubMed] [Google Scholar]
  29. Chun M. M., Jiang Y. (1998). Contextual cueing: implicit learning and memory of visual context guides spatial attention. Cogn. Psychol. 36, 28–71 10.1006/cogp.1998.0681 [DOI] [PubMed] [Google Scholar]
  30. Chun M. M., Jiang Y. (1999). Top-down attentional guidance based on implicit learning of visual covariation. Psychol. Sci. 10, 360–365 10.1111/1467-9280.00168 [DOI] [Google Scholar]
  31. Colgin L. L. (2011). Oscillations and hippocampal-prefrontal synchrony. Curr. Opin. Neurobiol. 21, 467–474 10.1016/j.conb.2011.04.006 [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Costello P., Jiang Y., Baartman B., McGlennen K., He S. (2009). Semantic and subword priming during binocular suppression. Conscious. Cogn. 18, 375–382 10.1016/j.concog.2009.02.003 [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Davenport J. L., Potter M. C. (2004). Scene consistency in object and background perception. Psychol. Sci. 15, 559–564 10.1111/j.0956-7976.2004.00719.x [DOI] [PubMed] [Google Scholar]
  34. Dehaene S., Changeux J. (2011). Experimental and theoretical approaches to conscious processing. Neuron 70, 200–227 10.1016/j.neuron.2011.03.018 [DOI] [PubMed] [Google Scholar]
  35. den Ouden H. E. M., Daunizeau J., Roiser J., Friston K. J., Stephan K. E. (2010). Striatal prediction error modulates cortical coupling. J. Neurosci. 30, 3210–3219 10.1523/JNEUROSCI.4458-09.2010 [DOI] [PMC free article] [PubMed] [Google Scholar]
  36. Denison R. N., Piazza E. A., Silver M. A. (2011). Predictive context influences perceptual selection during binocular rivalry. Front. Hum. Neurosci. 5:166. 10.3389/fnhum.2011.00166 [DOI] [PMC free article] [PubMed] [Google Scholar]
  37. Dux P. E., Visser T. A. W., Goodhew S. C., Lipp O. V. (2010). Delayed reentrant processing impairs visual awareness: an object substitution masking study. Psychol. Sci. 21, 1242–1247 10.1177/0956797610379866 [DOI] [PubMed] [Google Scholar]
  38. Eger E., Henson R. N., Driver J., Dolan R. J. (2007). Mechanisms of top-down facilitation in perception of visual objects studied by FMRI. Cereb. Cortex 17, 2123–2133 10.1093/cercor/bhl119 [DOI] [PMC free article] [PubMed] [Google Scholar]
  39. Egner T., Monti J. M., Summerfield C. (2010). Expectation and surprise determine neural population responses in the ventral visual stream. J. Neurosci. 30, 16601–16608 10.1523/JNEUROSCI.2770-10.2010 [DOI] [PMC free article] [PubMed] [Google Scholar]
  40. Esterman M., Yantis S. (2010). Perceptual expectation evokes category-selective cortical activity. Cereb. Cortex 20, 1245–1253 10.1093/cercor/bhp188 [DOI] [PMC free article] [PubMed] [Google Scholar]
  41. Fahle M. (1982). Binocular rivalry: suppression depends on orientation and spatial frequency. Vision Res. 22, 787–800 10.1016/0042-6989(82)90010-4 [DOI] [PubMed] [Google Scholar]
  42. Fahrenfort J. J., Scholte H. S., Lame V. A. F. (2007). Masking disrupts reentrant processing in human visual cortex. J. Cogn. Neurosci. 30, 7414–7422 [DOI] [PubMed] [Google Scholar]
  43. Felleman D. J., Van Essen D. C. (1991). Distributed hierarchical processing in the primate cortex. Cereb. Cortex 1, 1–47 10.1093/cercor/1.1.1 [DOI] [PubMed] [Google Scholar]
  44. Friedman A. (1979). Framing pictures: the role of knowledge in automatized encoding and memory for gist. J. Exp. Psychol. Gen. 108, 316–355 10.1037/0096-3445.108.3.316 [DOI] [PubMed] [Google Scholar]
  45. Friston K. J. (2005). A theory of cortical responses. Philos. Trans. R. Soc. Lond. B Biol. Sci. 360, 815–836 10.1098/rstb.2005.1622 [DOI] [PMC free article] [PubMed] [Google Scholar]
  46. Friston K. J. (2009). The free-energy principle: a rough guide to the brain? Trends Cogn. Sci. (Regul. Ed.) 13, 293–301 10.1016/j.tics.2009.04.005 [DOI] [PubMed] [Google Scholar]
  47. Friston K. J., Kiebel S. (2009). Predictive coding under the free-energy principle. Philos. Trans. R. Soc. Lond. B Biol. Sci. 364, 1211–1121 10.1098/rstb.2008.0300 [DOI] [PMC free article] [PubMed] [Google Scholar]
  48. Gamond L., George N., Lemaréchal J., Hugueville L., Adam C., Tallon-Baudry C. (2011). Early influence of prior experience on face perception. Neuroimage 54, 1415–1426 10.1016/j.neuroimage.2010.08.081 [DOI] [PubMed] [Google Scholar]
  49. Goolkasian P., Woodberry C. (2010). Priming effects with ambiguous figures. Atten. Percept. Psychophys. 72, 168–178 10.3758/APP.72.1.168 [DOI] [PubMed] [Google Scholar]
  50. Gorlin S., Meng M., Sharma J., Sugihara H., Sur M., Sinha P. (2012). Imaging prior information in the brain. Proc. Natl. Acad. Sci. U.S.A. 109, 1935–7940 10.1073/pnas.1111224109 [DOI] [PMC free article] [PubMed] [Google Scholar]
  51. Grill-Spector K., Kourtzi Z., Kanwisher N. (2001). The lateral occipital complex and its role in object recognition. Vision Res. 41, 1409–1422 10.1016/S0042-6989(01)00073-6 [DOI] [PubMed] [Google Scholar]
  52. Gronau N., Neta M., Bar M. (2008). Integrated contextual representation for objects’ identities and their locations. J. Cogn. Neurosci. 20, 371–388 10.1162/jocn.2008.20027 [DOI] [PubMed] [Google Scholar]
  53. Hohwy J. (2012). Attention and conscious perception in the hypothesis testing brain. Front. Psychol. 3:96. 10.3389/fpsyg.2012.00096 [DOI] [PMC free article] [PubMed] [Google Scholar]
  54. Hohwy J., Roepstorff A., Friston K. J. (2008). Predictive coding explains binocular rivalry: an epistemological review. Cognition 108, 687–701 10.1016/j.cognition.2008.05.010 [DOI] [PubMed] [Google Scholar]
  55. Hsieh P. J., Vul E., Kanwisher N. (2010). Recognition alters the spatial pattern of FMRI activation in early retinotopic cortex. J. Neurophysiol. 103, 1501–1507 10.1152/jn.00812.2009 [DOI] [PMC free article] [PubMed] [Google Scholar]
  56. Kaplan I. T., Metlay W. (1964). Light intensity and binocular rivalry. J. Exp. Psychol. 67, 22–26 10.1037/h0041379 [DOI] [PubMed] [Google Scholar]
  57. Kim R., Seitz A., Feenstra H., Shams L. (2009). Testing assumptions of statistical learning: is it long-term and implicit? Neurosci. Lett. 461, 145–149 10.1016/j.neulet.2009.06.030 [DOI] [PubMed] [Google Scholar]
  58. Kveraga K., Boshyan J., Bar M. (2007). Magnocellular projections as the trigger of top-down facilitation in recognition. J. Neurosci. 27, 13232–13240 10.1523/JNEUROSCI.3481-07.2007 [DOI] [PMC free article] [PubMed] [Google Scholar]
  59. Kveraga K., Ghuman A. S., Kassam K. S., Aminoff E., Hämäläinen M. S., Chaumon M., et al. (2011). Early onset of neural synchronization in the contextual associations network. Proc. Natl. Acad. Sci. U.S.A. 108, 3389–3394 10.1073/pnas.1013760108 [DOI] [PMC free article] [PubMed] [Google Scholar]
  60. Lamme V. A. F. (2010). How neuroscience will change our view on consciousness. Cogn. Neurosci. 1, 204–220 10.1080/17588921003731586 [DOI] [PubMed] [Google Scholar]
  61. Lamme V. A. F., Zipser K., Spekreijse H. (1998). Figure-ground activity in primary visual cortex is suppressed by anesthesia. Proc. Natl. Acad. Sci. U.S.A. 95, 3263–3268 10.1073/pnas.95.6.3263 [DOI] [PMC free article] [PubMed] [Google Scholar]
  62. Lamme V. A. F., Zipser K., Spekreijse H. (2002). Masking interrupts figure-ground signals in V1. J. Cogn. Neurosci. 14, 1044–1053 10.1162/089892902320474490 [DOI] [PubMed] [Google Scholar]
  63. Leopold D. A., Logothetis N. K. (1999). Multistable phenomena: changing views in perception. Trends Cogn. Sci. (Regul. Ed.) 3, 254–264 10.1016/S1364-6613(99)01332-7 [DOI] [PubMed] [Google Scholar]
  64. Livingstone M., Hubel D. (1988). Segregation of form, color, movement, and depth: anatomy, physiology, and perception. Science 240, 740–749 10.1126/science.3283936 [DOI] [PubMed] [Google Scholar]
  65. MacEvoy S. P., Epstein R. A. (2011). Constructing scenes from objects in human occipitotemporal cortex. Nat. Neurosci. 14, 1323–1329 10.1038/nn.2903 [DOI] [PMC free article] [PubMed] [Google Scholar]
  66. Malach R., Reppas J. B., Benson R. R., Kwong K. K., Jiang H., Kennedy W. A., et al. (1995). Objected related activity revealed by functional magnetic resonance imaging in human occipital cortex. Proc. Natl. Acad. Sci. U.S.A. 92, 8135–8139 10.1073/pnas.92.18.8135 [DOI] [PMC free article] [PubMed] [Google Scholar]
  67. Maunsell J. H. R., Nealey T. A., DePriest D. D. (1990). Magnocellular and parvocellular contributions to responses in the middle temporal visual area (MT) of the macaque monkey. J. Neurosci. 10, 3323–3334 [DOI] [PMC free article] [PubMed] [Google Scholar]
  68. Melloni L., Schwiedrzik C. M., Müller N., Rodriguez E., Singer W. (2011). Expectations change the signatures and timing of electrophysiological correlates of perceptual awareness. J. Neurosci. 31, 1386–1396 10.1523/JNEUROSCI.4570-10.2011 [DOI] [PMC free article] [PubMed] [Google Scholar]
  69. Merigan W. H., Maunsell J. H. R. (1993). How parallel are the primate visual pathways? Annu. Rev. Neurosci. 16, 369–402 10.1146/annurev.neuro.16.1.369 [DOI] [PubMed] [Google Scholar]
  70. Meyer K. (2012). Another remembered present. Science 335, 415–416 10.1126/science.1214449 [DOI] [PubMed] [Google Scholar]
  71. Mumford D. (1992). On the computational architecture of the neocortex. Biol. Cybern. 251, 241–251 10.1007/BF00198477 [DOI] [PubMed] [Google Scholar]
  72. Murray E. A., Mishkin M. (1998). Object recognition and location memory in monkeys with excitotoxic lesions of the amygdala and hippocampus. J. Neurosci. 18, 6568–6582 [DOI] [PMC free article] [PubMed] [Google Scholar]
  73. Nyberg L., Tulving E., Habib R., Nilsson L. G., Kapur S., Houle S., et al. (1995). Functional brain maps of retrieval mode and recovery of episodic information. Neuroreport 7, 249–252 10.1097/00001756-199512290-00060 [DOI] [PubMed] [Google Scholar]
  74. Oliva A., Torralba A. (2007). The role of context in object recognition. Trends Cogn. Sci. (Regul. Ed.) 11, 520–527 10.1016/j.tics.2007.09.009 [DOI] [PubMed] [Google Scholar]
  75. Ooi T. L., He Z. J. (1999). Binocular rivalry and visual awareness: the role of attention. Perception 28, 551–574 10.1068/p2923 [DOI] [PubMed] [Google Scholar]
  76. Palmer E. (1975). The effect of contextual scenes on the identification of objects. Mem. Cognit. 3, 519–526 10.3758/BF03197524 [DOI] [PubMed] [Google Scholar]
  77. Pascual-Leone A., Walsh V. (2001). Fast backprojections from the motion to the primary visual area necessary for visual awareness. Science 292, 510–512 10.1126/science.1057099 [DOI] [PubMed] [Google Scholar]
  78. Pearson J., Clifford C., Tong F. (2008). The functional impact of mental imagery on conscious perception. Curr. Biol. 18, 982–986 10.1016/j.cub.2008.05.048 [DOI] [PMC free article] [PubMed] [Google Scholar]
  79. Peters J., Daum I., Gizewski E., Forsting M., Suchan B. (2009). Association evoked during memory encoding recruit the context network. Hippocampus 19, 141–151 10.1002/hipo.20490 [DOI] [PubMed] [Google Scholar]
  80. Piefke M., Weiss P. H., Zilles K., Markowitsch H. J., Fink G. R. (2003). Differential remoteness and emotional tone modulate the neural correlates of autobiographical memory. Brain 126, 650–668 10.1093/brain/awg064 [DOI] [PubMed] [Google Scholar]
  81. Rao R. P., Ballard D. H. (1999). Predictive coding in the visual cortex: a functional interpretation of some extra-classical receptive-field effects. Nat. Neurosci. 2, 79–87 10.1038/4580 [DOI] [PubMed] [Google Scholar]
  82. Rauchs G., Orban P., Balteau E., Schmidt C., Degueldre C., Luxen A., et al. (2008). Partially segregated neural networks for spatial and contextual memory in virtual navigation. Hippocampus 18, 503–518 10.1002/hipo.20411 [DOI] [PubMed] [Google Scholar]
  83. Rempel-Clower N. L., Barbas H. (2000). The laminar pattern of connections between prefrontal and anterior temporal cortices in the Rhesus monkey is related to cortical structure and function. Cereb. Cortex 10, 851–865 10.1093/cercor/10.9.851 [DOI] [PubMed] [Google Scholar]
  84. Reynolds R. I. (1985). The role of object-hypothesis in the organization of fragmented figures. Perception 14, 49–52 10.1068/p140049 [DOI] [PubMed] [Google Scholar]
  85. Ro T., Breitmeyer B., Burton P., Singhal N. S., Lane D. (2003). Feedback contributions to visual awareness in human occipital cortex. Curr. Biol. 13, 1038–1041 10.1016/S0960-9822(03)00337-3 [DOI] [PubMed] [Google Scholar]
  86. Rosch E., Mervis C. B., Gray W. D., Johnson D. M., Boyes-Braem P. (1976). Basic objects in natural categories. Cogn. Psychol. 439, 382–439 10.1016/0010-0285(76)90013-X [DOI] [Google Scholar]
  87. Sachs O., Weis S., Zellagui N., Sass K., Huber W., Zvyagintsev M., et al. (2011). How different types of conceptual relations modulate brain activation during semantic priming. J. Cogn. Neurosci. 23, 1263–1273 10.1162/jocn.2010.21483 [DOI] [PubMed] [Google Scholar]
  88. Schyns P. G., Oliva A. (1994). From blobs to boundaries: evidence for time- and spatial-scale-dependent scene recognition. Psychol. Sci. 5, 195–201 10.1111/j.1467-9280.1994.tb00500.x [DOI] [Google Scholar]
  89. Shapley R. (1990). Visual sensitivity and parallel retinocortical channels. Annu. Rev. Psychol. 41, 635–658 10.1146/annurev.ps.41.020190.003223 [DOI] [PubMed] [Google Scholar]
  90. Sotiropoulos G., Seitz A. R., Seriès P. (2011). Changing expectations about speed alters perceived motion direction. Curr. Biol. 21, 883–884 10.1016/j.cub.2011.03.070 [DOI] [PubMed] [Google Scholar]
  91. Squire L. R., Stark C. E. L., Clark R. E. (2004). The medial temporal lobe. Annu. Rev. Neurosci. 27, 279–306 10.1146/annurev.neuro.27.070203.144130 [DOI] [PubMed] [Google Scholar]
  92. Steinman B. A., Steinman S. B. (1997). Transient visual attention is dominated by the magnocellular stream. Vision Res. 37, 17–23 10.1016/S0042-6989(96)00151-4 [DOI] [PubMed] [Google Scholar]
  93. Sterzer P., Frith C., Petrovic P. (2008). Believing is seeing: expectations alter visual awareness. Curr. Biol. 18, 697–698 10.1016/j.cub.2008.06.021 [DOI] [PubMed] [Google Scholar]
  94. Sterzer P., Kleinschmidt A., Rees G. (2009). The neural bases of multistable perception. Trends Cogn. Sci. (Regul. Ed.) 13, 310–318 10.1016/j.tics.2009.04.006 [DOI] [PubMed] [Google Scholar]
  95. Stocker A. A., Simoncelli C. P. (2006). Noise characteristics and prior expectations in human visual speed perception. Nat. Neurosci. 9, 578–585 10.1038/nn1669 [DOI] [PubMed] [Google Scholar]
  96. Summerfield C., Egner T. (2009). Expectation (and attention) in visual cognition. Trends Cogn. Sci. (Regul. Ed.) 13, 403–409 10.1016/j.tics.2009.06.003 [DOI] [PubMed] [Google Scholar]
  97. Summerfield C., Egner T., Greene M., Koechlin E., Mangels J., Hirsch J. (2006). Predictive codes for forthcoming perception in the frontal cortex. Science 314, 1311–1314 10.1126/science.1132028 [DOI] [PubMed] [Google Scholar]
  98. Sun J., Perona P. (1998). Where is the sun? Nat. Neurosci. 1, 183–184 10.1038/630 [DOI] [PubMed] [Google Scholar]
  99. Supèr H., Spekreijse H., Lamme V. A. F. (2001). Two distinct modes of sensory processing observed in monkey primary visual cortex (V1). Nat. Neurosci. 4, 304–310 10.1038/85170 [DOI] [PubMed] [Google Scholar]
  100. Szpunar K. K., Chan J. C. K., McDermott K. B. (2009). Contextual processing in episodic future thought. Cereb. Cortex 19, 1539–1548 10.1093/cercor/bhn191 [DOI] [PubMed] [Google Scholar]
  101. Thorpe S., Fize D., Marlot C. (1996). Speed of processing in the human visual system. Nature 381, 520–522 10.1038/381520a0 [DOI] [PubMed] [Google Scholar]
  102. Tononi G., Edelman G. E. (1998). Consciousness and Complexity. Science 282, 1846–1851 10.1126/science.282.5395.1846 [DOI] [PubMed] [Google Scholar]
  103. Tsuchiya N., Koch C. (2005). Continuous flash suppression reduces negative afterimages. Nat. Neurosci. 8, 1096–1101 10.1038/nn1500 [DOI] [PubMed] [Google Scholar]
  104. Tsukiura T., Cabeza R. (2008). Orbitofrontal and hippocampal contributions to memory for face-name associations: the rewarding power of a smile. Neuropsychologia 46, 2310–2319 10.1016/j.neuropsychologia.2008.03.013 [DOI] [PMC free article] [PubMed] [Google Scholar]
  105. Turk-Browne N. B., Scholl B. J., Johnson M. K., Chun M. M. (2010). Implicit perceptual anticipation triggered by statistical learning. J. Neurosci. 30, 11177–11187 10.1523/JNEUROSCI.0858-10.2010 [DOI] [PMC free article] [PubMed] [Google Scholar]
  106. Ullman S. (1995). Sequence seeking and counter streams: a computational model for bidirectional information flow in the visual cortex. Cereb. Cortex 5, 1–11 10.1093/cercor/5.1.1 [DOI] [PubMed] [Google Scholar]
  107. Webster M. J., Bachevalier J., Ungerleider L. G. (1994). Connections of inferior temporal areas TEO and TE with parietal and frontal cortex in macaque monkeys. Cereb. Cortex 4, 470–483 10.1093/cercor/4.5.470 [DOI] [PubMed] [Google Scholar]
  108. Wyart V., Nobre A. C., Summerfield C. (2012). Dissociable prior influences of signal probability and relevance on visual contrast sensitivity. Proc. Natl. Acad. Sci. U.S.A. 109, 3593–3598 10.1073/pnas.1120118109 [DOI] [PMC free article] [PubMed] [Google Scholar]
  109. Yeo B. T. T., Krienen F. M., Sepulcre J., Sabuncu M. R., Lashkari D., Hollinshead M., et al. (2011). The organization of the human cerebral cortex estimated by intrinsic functional connectivity. J. Neurophysiol. 106, 1125–1165 10.1152/jn.00338.2011 [DOI] [PMC free article] [PubMed] [Google Scholar]

Articles from Frontiers in Psychology are provided here courtesy of Frontiers Media SA

RESOURCES