Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2014 Sep 3.
Published in final edited form as: Psychol Sci. 2010 Mar 23;21(5):682–691. doi: 10.1177/0956797610366099

Conceptual Penetration of Visual Processing

Gary Lupyan 1,2, Sharon L Thompson-Schill 1,2,3, Daniel Swingley 1,3
PMCID: PMC4152984  NIHMSID: NIHMS527628  PMID: 20483847

Abstract

In traditional hierarchical models of information processing, visual representations feed into conceptual systems, but conceptual categories do not exert an influence on visual processing. We provide evidence, across four experiments, that conceptual information can in fact penetrate early visual processing, rather than merely biasing the output of perceptual systems. Participants performed physical-identity judgments on visually equidistant pairs of letter stimuli that were either in the same conceptual category (Bb) or in different categories (Bp). In the case of nonidentical letters, response times were longer when the stimuli were from the same conceptual category, but only when the letters were presented sequentially. The differences in effect size between simultaneous and sequential trials rules out a decision-level account. An additional experiment using animal silhouettes replicated the major effects found with letters. Thus, performance on an explicitly visual task was influenced by conceptual categories. This effect depended on processing time, immediately preceding experience, and stimulus typicality, which suggests that it was produced by the direct influence of category knowledge on perception, rather than by a postperceptual decision bias.

Keywords: visual processing, categorization, modularity of vision, top-down effects, perceptual learning, knowledge-based vision


Extracting meaning from a perceptual signal is an act of categorization, requiring the observer to highlight some stimulus features while overlooking (abstracting over) others that are irrelevant to a given category (Harnad, 2005). Humans categorize objects with incredible speed. By some measures, basic-level categorization occurs in parallel with object detection; by the time an object is detected, it is already categorized to some degree (Grill-Spector & Kanwisher, 2005; cf. Mack, Gauthier, Sadr, & Palmeri, 2008). When participants are monitoring for a known object category, neural signatures of categorization can be measured as early as 150 ms after stimulus onset (VanRullen & Thorpe, 2001; cf. Johnson & Olshausen, 2003). It has been argued that such rapid categorization can be achieved by biasing visual processing to favor low-level features diagnostic of the relevant category (Delorme, Rousselet, Mace, & Fabre-Thorpe, 2004; Johnson & Olshausen, 2003; McCotter, Gosselin, Maccabee, & Schyns, 2005; Schyns & Oliva, 1999).

In traditional hierarchical models of information processing, early visual processing feeds into conceptual systems, but the latter do not influence visual processes (Glezer, Jiang, & Riesenhuber, 2009; Pylyshyn, 1999; Riesenhuber & Poggio, 2000). Here, we provide evidence, from four experiments, that conceptual information can in fact penetrate early visual processing, rather than merely biasing the output of perceptual systems.

There is often a conflict between the physical and conceptual relationships between objects. Consider the letter pairs B-b and B-p. The physical relationship between the letters is identical in the two pairs—B is physically equidistant from p and b. However, conceptually, B is more similar to b than to p. A conceptual effect on perception—greater perceived similarity between B and b than between B and p—can arise in different ways.

One possibility, consistent with studies of perceptual learning (Goldstone, 1994, 1998; Goldstone, Steyvers, Spencer-Smith, & Kersten, 2004; Harnad, 1987; Kuhl, 1994; Livingston, Andrews, & Harnad, 1998; Notman, Sowden, & Ozgen, 2005; Schyns & Rodet, 1997), is that long-term experience categorizing B and b as members of the same category increases their perceptual similarity, and, conversely, that categorizing B and p into different categories decreases their similarity, through a gradual, bottom-up retuning of visual feature detectors. Empirical demonstrations of these effects typically involve perceptual comparisons of novel items before and after categorization experience.

However, changes in similarity structure can also result from on-line, transient modulation of perceptual processing by higher-level conceptual representations (a top-down process). Consistent with this view are studies showing that visual categorical perception can sometimes be disrupted by verbal interference (which arguably acts to block conceptual effects on perception; A. Gilbert, Regier, Kay, & Ivry, 2006; Winawer et al., 2007). Such a modulatory account is compatible with electrophysiological findings that show very rapid top-down modulation of early visual processing (e.g., V1) starting 10 to 50 ms after stimulus onset (Hupe et al., 2001; for reviews, see C. Gilbert & Sigman, 2007; Lamme & Roelfsema, 2000). Such rapid modulation is made possible by a surprisingly fast posterior-to-anterior activation flow, with visual cortex activating prefrontal cortex in as little as 30 ms (Foxe & Simpson, 2002). Foxe and Simpson argued that these results “provide a context for appreciating the 100–400 ms of processing necessary prior to response output in humans, [which leaves] ample time for multiple cortical interactions at all levels of the system” (p. 145).

Finally, some researchers have argued that apparent conceptual effects on perception do not involve alterations of visual processing at all, and can be fully explained by high-level decision biases (e.g., Pylyshyn, 1999).

In this article, we present four experiments aimed at testing the idea that conceptual effects on perceptual processing are best described in terms of a dynamic top-down process. We hypothesized that the perceived visual similarity between two stimuli will become increasingly affected by the category membership of the stimuli over the course of several hundred milliseconds, as the conceptual category information feeds back on lower-level perceptual representations.

Experiment 1

A highly reliable measure of visual similarity is the time it takes to determine whether two stimuli are physically different. Increases in perceived visual similarity have a behavioral signature of greater response times (RTs). Thus, effects of conceptual categories on perceptual processing can be studied by manipulating the conceptual relationship between stimuli while keeping the physical relationship constant. We define the category effect as the difference in RTs between within-conceptual-category pairs (Bb) and between-category pairs (Bp). We hypothesized that when the two stimuli were presented simultaneously, participants would respond before the conceptual representations exerted a significant influence on the ongoing visual processing. Thus, we predicted a small or nonexistent category effect for simultaneously presented stimuli. However, with greater time for processing and categorizing one of the items, its visual representation would become more similar to the representations of other members of the same category or more different from the representations of members of different categories. We thus contrasted the category effect during simultaneous judgments and sequential judgments. Our prediction was that it would take longer to respond “different” to within-category trials (Bb) than to between-category trials (Bp), and that the size of this category effect would be larger for sequentially presented than for simultaneously presented pairs.

Method

Twelve University of Pennsylvania undergraduates participated in Experiment 1 for course credit. They were asked to perform a speeded same/different task, responding via a keyboard button press to indicate “same” if the letters in a pair were physically identical, and “different” otherwise. All pairs were composed from the set of the letters B, b, and p. The critical pairs were the two different pairs: Bb (within category) and Bp (between category). Pixel by pixel, the letters B and b were as similar to each other as the letters B and p. Any differences in responding to these two pairs could not be due to differences in visual similarity, and thus would have to be attributed to conceptual differences (Lupyan, 2008a).

On half of the trials, the letters were presented simultaneously. On the remaining half, the first letter was displayed on the screen for 150, 300, 450, or 600 ms before the second letter appeared. The first letter remained on the screen during the variable delay (Fig. 1).1 To discourage the allocation of attention to a specific part of the display during the delay, we varied the location of the letters; on each trial, the two stimuli (~0.7° × 0.9°) appeared in two of four possible (randomly selected) locations equidistant (~1.5°) from a central fixation cross. There were no memory demands, as both stimuli remained on the screen until a response was made. The intertrial interval was 750 ms.

Fig. 1.

Fig. 1

Experimental design for Experiments 1 through 3. The two letters on each trial were the same, physically different but from the same conceptual category, or physically different and from different conceptual categories. These letters were presented either simultaneously or sequentially. On the sequential trials, the first stimulus remained visible during a variable delay (stimulus onset asynchrony, or SOA) before onset of the second stimulus.

Participants completed 12 practice trials followed by 576 experimental trials. There were equal numbers of same and different trials, and of within- and between-category trials. Participants were encouraged to respond as quickly as possible without compromising accuracy. A buzz sounded after incorrect responses. Hand-to-response assignment was counterbalanced across subjects.

Results and discussion

Errors were rare (< 5%). We analyzed RTs from the onset of the second stimulus. Trials responded to incorrectly were excluded from analysis, as were trials with RTs shorter than 200 ms or longer than 1,200 ms (3.3%). Hand-to-response assignment did not interact with any variables of interest (Fs < 1).

RTs were analyzed using repeated measures analyses of variance (ANOVAs). Participants responded more quickly to same trials (M = 550 ms) than to different trials (M = 577 ms), F(1, 11) = 10.16, p = .009—an example of a previously observed “fast-same” effect (Posner, 1978). Unsurprisingly, the extra processing time available in sequential trials yielded faster RTs than were observed in simultaneous trials, F(1, 11) = 26.70, p < .0005.

The critical prediction concerned the different trials. We predicted that RTs on these trials would be faster when the two stimuli were categorically different (Bp) than when they were categorically identical (Bb), but only when the first stimulus received extra processing time (i.e., when there was a nonzero stimulus onset asynchrony, or SOA). RTs were significantly slower on within-category trials (M = 589 ms) than on between-category trials (M = 565 ms), F(1, 11) = 9.91, p = .009. Critically, there was a significant interaction between category type (same vs. different) and SOA (zero vs. nonzero), F(1, 11) = 7.99, p = .016 (Fig. 2a). When the two letters were presented simultaneously, there was an unreliable 11-ms difference between the two trial types, t(11) = 1.78, p > .1 (two-tailed). In contrast, the 38-ms difference observed on the sequential trials was significant, t(11) = 3.32, p = .007. There was no evidence of speed-accuracy trade-offs. Participants were not only faster, but also slightly (1%) more accurate on the between-category trials, F(1, 11) = 2.98, p = .11. There were no effects of presentation order (e.g., Bp vs. pB), t < 1.

Fig. 2.

Fig. 2

Results from (a) Experiment 1 (upright letters) and (b) Experiment 2 (rotated letters). The graphs show mean reaction times on different trials as a function of stimulus onset asynchrony (SOA), separately for within-category and between-category trials. Error bars represent ±1 SE of the mean difference score for a given SOA.

The magnitude of the category effect did not vary reliably among the nonzero SOAs, but as Figure 2a illustrates, the effect became more reliable as SOA increased. Simultaneous Levine tests showed that variance was significantly higher at the 150-ms SOA than at the other delays and decreased from the 300-ms to the 450-ms SOA, ps < .01.

It may at first appear that these results are compatible with two accounts: a conceptual-penetration account, in which categories affect visual representations, and a decision-level account, in which same-category pairs (e.g., Bb) introduce a "same" response bias that competes with the desired "different" response. According to the latter account, the decision to respond "same" or "different" is a linear combination of two input channels, one used to make the decision on the basis of visual features and one used to make the decision on the basis of conceptual features (Hawkins et al., 1990); the existence (and size) of the category effect would then depend on the weights assigned to the channels. However, in the present paradigm, neither the visual nor the conceptual response alternatives can compete until both stimuli are visible (i.e., the channel weights are effectively 0 until both stimuli are present). Thus, a decision-level account predicts a category effect that is independent of SOA (see A Test of the Decision-Level Model Account in the Supplemental Material available online). In contrast, according to the conceptual-penetration account, categorization of the first stimulus influences the perceptual processing of the second, and the feedback effect of categories on perceptual processing is more limited when the stimuli are presented simultaneously or with a short SOA.

Experiment 2

We hypothesized that conceptual effects on perceptual processing are modulated by the strength of the association between the visual form and the conceptual category (Lupyan, 2007, 2008a). Manipulations that disrupt this association are therefore predicted to reduce or eliminate the conceptual-penetration effect (the RT difference between Bb and Bp trials). In Experiment 2, we achieved such disruption by repeating the procedures of Experiment 1 with the stimuli rotated 90°. Rotating the stimuli preserved the physical relationship between them, but we assumed it would partially disrupt the association between the visual forms and the conceptual categories. Fifteen University of Pennsylvania undergraduates participated for course credit.

Trials responded to incorrectly were excluded from analysis, as were trials with RTs shorter than 200 ms or longer than 1,200 ms (1.0%). In contrast to Experiment 1, which showed a large effect of category type on sequential trials, Experiment 2 showed no evidence of a category effect, as confirmed by a repeated measures ANOVA with category effect as a fixed factor F(1, 14) < 1 (Fig. 2b). Accuracy was high (M = 96.5%) and did not significantly differ between conditions; the direction of accuracy differences mirrored the direction of RT analyses.

To compare Experiments 1 and 2 directly, we tested the three-way Experiment × SOA (zero vs. nonzero) × Category Type interaction, which proved to be significant, F(1, 25) = 4.03 p = .048. Further analysis revealed that the Experiment × Category Type interaction was significant for the sequential trials, F(1, 25) = 7.10, p = .010, but not for the simultaneous trials, F < 1. That is, although responses to the sequential between-category (Bp) trials were significantly faster than responses to the sequential within-category (Bb) trials in Experiment 1 (Mbetween = 535 ms, Mwithin = 573 ms), this pattern did not hold for sequential trials in Experiment 2 (Mbetween = 577 ms, Mwithin = 580 ms).

These results confirm the prediction that manipulations that maintain all the low-level visual components of the stimuli, but disrupt their relation to the conceptual category, are sufficient to disrupt conceptual effects on visual processing.

Experiment 3

We reasoned that just as disrupting the association between the visual form and its category decreased the influence of the conceptual category on visual processing, strengthening the association would increase the strength or speed of the category effect.

Method

This experiment was identical to Experiment 1 except that the main task was preceded by a 5-min overt categorization task in which participants were instructed to respond with the “b” key to Bs and bs and with the “p” key to ps. On each trial, one letter appeared by itself randomly in one of the four positions used in Experiment 1 and remained on the screen until a response. There were 80 b trials and 80 p trials randomly intermixed. Following this categorization task, the same/different judgment trials were presented as in Experiment 1. Nineteen University of Pennsylvania undergraduates participated for course credit.

Results and discussion

Mean RT for the overt classification phase was 492 ms (decreasing from 532 ms for the first 40 trials to 472 ms for the last 40). Classification errors were rare (< 5%). As in Experiment 1, the RT analysis of the same/different task excluded incorrect responses and trials with RTs shorter than 200 ms or longer than 1,200 ms (2.2%).

A repeated measures ANOVA conducted on the different trials in the main task revealed significant effects of SOA (zero vs. nonzero), F(1, 18) = 72.89, p < .0005, and category type (within- vs. between-category trials), F(1, 18) = 16.55, p = .001. Unlike in Experiment 1, there was no interaction between these factors, F < 1 (Fig. 3).

Fig. 3.

Fig. 3

Mean reaction times on different trials in Experiment 3 as a function of stimulus onset asynchrony (SOA), separately for within-category and between-category trials. Error bars represent ±1 SE of the mean difference score for a given SOA.

Planned t tests (two-tailed) revealed that, unlike in Experiment 1, there was a significant category effect when the stimuli were presented simultaneously (M = 18 ms), t(18) = 4.62, p < .0005. The effect at the 150-ms SOA was of comparable magnitude to that observed in Experiment 1 (~31 ms), but in this case was reliably greater than 0, t(18) = 2.43, p = .026. Unlike in Experiment 1, there was a reliable main effect of category type on accuracy. Accuracy was significantly greater for the between-category trials (M = 96.3%) than for the within-category trials (M = 93.2%), F(1, 18) = 19.71, p < .0005. There was no interaction between category type and SOA, F < 1. The effect of category type on accuracy was greater in Experiment 3 than in Experiment 1, F(1, 29) = 4.29, p = .039.

In an unexpected departure from Experiment 1, RTs for within-category trials were no greater than RTs for between-category trials for SOAs of 450 and 600 ms. For these longer SOAs, the category effect was significantly greater in Experiment 1 than in Experiment 3, F(1, 29) = 9.46, p = .005. This result provides an important constraint on the proposed account, suggesting that the category effect not only has a dynamic onset, but also has an offset. Strengthening the association between the visual form and the conceptual category appears to shift the timing of the category effect. We tested this prediction by identifying the time point at which the category effect was largest for each subject and comparing the means of the Gaussian fits between Experiments 1 and 3 (Fig. 4). A two-tailed t test showed the peak effect timing to be significantly earlier for Experiment 3 (245 ms) than for Experiment 1 (375 ms), t(29) = 2.19, p = .036.

Fig. 4.

Fig. 4

Probability density functions of participants in Experiments 1 and 3. Each point represents the normalized frequency of a peak category effect (reaction time for within-category trials minus reaction time for across-category trials) for a particular stimulus onset asynchrony (SOA).

In addition to comparing the timing of the category effect between experiments, we examined whether subjects who were faster in categorizing the stimuli during the pretask categorization trials demonstrated an earlier category effect in the main task. Categorization RTs did not predict the overall size of the category effect, r = −.05, but did predict the category effect for simultaneous judgments, r = −.755, p < .001. The most parsimonious account of these results is that faster categorization induces an earlier, but not necessarily larger, category effect.

To summarize, a brief “training” session in which participants categorized (already highly overlearned) letter stimuli caused category knowledge to affect simple same/different judgments more quickly, such that the effect was observed even at SOAs of 0. Moreover, the fastest categorizers were the most likely to show a reliable category effect at an SOA of 0. Category training also led to reduced accuracy on within-category trials relative to between-category trials.

Rather than changing the overall magnitude of the category effect, classification training appeared to shift the category effect earlier in time. Although a decision-level account also predicts that classification training would produce a more robust category effect (insofar as the decision based on category identity would compete more effectively with a decision based on physical identity), a decision-level account does not predict that categorization would shift the onset and offset of the effect, because, as mentioned earlier, category-level competition cannot begin until both stimuli are visible.

Experiment 4

The purpose of Experiment 4 was to generalize the conceptual-penetration effect to a richer set of stimuli. These richer stimuli allowed us to introduce greater perceptual variability and examine effects of stimulus typicality. Insofar as more typical stimuli are categorized more quickly and more reliably, processing a typical stimulus should lead to a stronger category effect than processing an atypical stimulus.

Method

Twenty-one subjects participated in Experiment 4 for course credit. One was excluded because of a high error rate (2.77 SD above the mean). The design was identical to that of Experiment 1 except that the letters were replaced by richer stimuli: silhouettes of cats and dogs (Quinn, Eimas, & Tarr, 2001; see Fig. 5). These images have the advantage of being meaningful and easily classified and of forcing participants to rely on global shape information rather than local image properties. To further increase variability, we manipulated the orientation of the silhouettes from trial to trial, although on any given trial, both animals faced to the right or to the left. Each participant completed 408 trials. All stimulus pairs were balanced in a 2 (same vs. different) × 2 (within vs. between category) × 2 (simultaneous vs. sequential) design. A separate group of 6 participants performed 144 trials of speeded classification (analogous to the first part of Experiment 3). These data were used to classify the stimuli into typical and atypical exemplars, in order to predict the size of the category effect for different combinations of stimuli.

Fig. 5.

Fig. 5

The stimuli used in Experiment 4 and the mean time to classify each as a “cat” or a “dog.” C1 and D1 served as “typical” stimuli. C3 and D3 served as “atypical” stimuli.

Results and discussion

Errors were rare (< 5%). Trials responded to incorrectly were excluded from analysis, as were trials with RTs shorter than 200 ms or longer than 1,500 ms (3.9%). RTs were faster for same trials (M = 727 ms) than for different trials (M = 763 ms), F(1, 19) = 11.27, p = .003. For the different trials, participants were always slower to respond to cata-catb trials (M = 814 ms) than to cat-dog trials (M = 747 ms), F(1, 19) = 61.36, p < .0005. This difference remained relatively constant for all SOAs, which suggests that the cats were more physically similar to each other than to dogs (i.e., conceptual similarity was confounded with physical similarity).

In contrast, the physical similarity between two dogs was not greater than the physical similarity between cats and dogs. Participants were as fast to respond “different” on doga-dogb trials as on dog-cat trials, but only when the two pictures appeared simultaneously (Fig. 6). When stimuli were presented sequentially, RTs for between-category trials became faster than RTs for within-category trials, resulting in a significant Category Type × SOA interaction, as in Experiment 1, F(1, 19) = 6.59, p = .019. Planned comparisons did not reveal a significant category effect for simultaneous trials, F < 1, and revealed a marginal effect for sequential trials, F(1, 19) = 3.95, p = .061. As is evident in Figure 6, the effect was largest for the 300-ms SOA (Mwithin-category = 745 ms, Mbetween-category = 695 ms), at which point the difference was significantly different from 0, F(1, 19) = 4.70, p = .043. As in Experiment 3, the effect declined for longer SOAs.

Fig. 6.

Fig. 6

Mean reaction times on different trials in Experiment 4 as a function of stimulus onset asynchrony (SOA), separately for within-category (cata-catb, doga-dogb) and between-category (cat-dog) trials. Error bars represent ±1 SE of the mean difference score for a given SOA.

The larger diversity of stimuli in this experiment allowed us to examine effects of category strength more directly than in the previous experiments. A more rapidly categorized stimulus is predicted to influence subsequent perceptual processing more than a more slowly categorized stimulus. Therefore, we predicted that seeing a more typical category member as the first stimulus in a sequential trial would produce a greater category effect than seeing a less typical stimulus first. To test this prediction, we used the classification RTs from the 6 additional subjects to select the cats and dogs with the fastest (Fig. 5: C1, D1) and slowest (Fig. 5: C3, D3) classification RTs. We then contrasted two types of within-category different trials: typical → atypical trials (C1 → C3 or D1 → D3) and atypical → typical trials (C3 → C1 or D3 → D1). As predicted, typical → atypical trials yielded significantly longer RTs (M = 797 ms) than atypical → typical trials (M = 711 ms), F(1, 19) = 10.49, p = .004. Recall that the stimuli being judged in the two cases were exactly the same.

Finally, we examined typicality effects as a function of presentation time. Order of presentation is undefined for simultaneous trials, but we could compare the size of the category effect for simultaneous trials and sequential trials, the latter being divided into those that presented a typical cat or dog as the first stimulus (typical first) and those that did not (atypical first). There was a significant Typicality Condition × Category Type interaction: The category effect was larger for the typical-first than for the atypical-first trials, F(1, 19) = 5.29, p = .03. Moreover, the category effect increased from simultaneous to sequential trials for the typical-first trials, t(19) = 2.94, p = .008, but not the atypical-first trials, t(19) < 1.

Experiment 4 extended the main results of the earlier studies to a more diverse set of stimuli. With natural categories it is normal for within-category physical similarity to be greater than between-category physical similarity, and this was the case with cata-catb versus cat-dog pairs. For these pairs, the magnitude of the category effect did not change with SOA and can be wholly explained through differences in physical similarity (Lupyan, 2008a). In contrast, RTs to doga-dogb pairs were no different from RTs to cat-dog pairs in the case of simultaneous presentations, but diverged for longer SOAs. The degree of this divergence was mediated by relative classification difficulty (measured by data collected from a separate group of participants). The category effect was larger when the more typical (easier to classify) stimulus was presented first. These results support the conclusion that category knowledge affects even very simple perceptual judgments, and that factors such as typicality and task timing affect this process in predictable ways (Lupyan, 2007, 2008b).

General Discussion

The time subjects took to make a simple visual decision about two familiar pictures was influenced by the categorical relationship between the pictures. Judging two pictures to be physically “different” took longer when they were in the same category than when they belonged to different categories, even though the pictures were visually equidistant in the two cases (Experiments 1 and 3). This category effect disappeared when the pictures were made less meaningful through a 90° rotation (Experiment 2). Critically, the category effect was absent when the two pictures were presented simultaneously, and emerged over the course of a delay between the presentation of the first and second stimulus (Experiments 1 and 4). Providing a brief categorization session prior to the main task shifted the onset of the category effect, producing a reliable category effect even for simultaneously presented stimuli (Experiment 3). We also found that subjects who were the fastest categorizers showed the most reliable category effect on simultaneous trials. Experiment 4 extended these findings in two important ways. First, we generalized the results to a new stimulus set. Second, we showed that the size of the category effect varies as a function of typicality: The time to judge two stimuli from the same category as different is longer when the typical stimulus is presented first.

These results provide behavioral evidence of conceptual penetration of visual processing, supporting the notion of categorical perception as a dynamic process, arising from a modulation of visual representations by higher-level conceptual representations (an account consistent with the findings of A. Gilbert et al., 2006; A. Gilbert, Regier, Kay, & Ivry, 2008; Winawer et al., 2007; see McMurray, Aslin, Tanenhaus, Spivey, & Subik, 2008, for an extension to the speech domain). Unlike demonstrations of task-demands on visual processing (e.g., Schyns & Oliva, 1999), the present results show that visual processing in the service of the very same task (physical same/different judgments) is affected by non-visual properties—the conceptual relationship between the stimuli that are being evaluated.

Although the present studies cannot establish the neural locus of the category effect, the results rule out a simple decision-level model according to which conceptual categories influence a global decision process without influencing visual representations (Mitterer, Horschig, Musseler, & Majid, 2009; Norris, McQueen, & Cutler, 2000). This model is ruled out because the visual and conceptual channels can compete only after both stimuli are visible, which means that SOA should not affect the size of the category effect. Such a model also does not predict the typicality asymmetries in Experiment 4.

Consider the finding that participants take longer to respond “different” to two dogs when a typical dog is followed by an atypical dog than when the order of the same stimuli is reversed (Experiment 4). Clearly, a response requires processing of both stimuli. According to standard accounts, the representations of the two stimuli are compared, and the speed of the decision depends on the distance between the two stimuli in a neural state space (which might reflect visual or conceptual dimensions, or both). If the representations are conceived of as static locations in state space, then, counter to the present results, there should be no asymmetry between typical-first and atypical-first trials (, i.e., distance[typical,atypical] = distance[typical,atypical]). In contrast, according to the conceptual-penetration account, seeing a typical dog activates a conceptual category that alters the visual processing of other dogs. This resulting increase in similarity between the first and second dogs produces a longer latency for a “different” response. Seeing the atypical dog first invokes this same process, but to a lesser degree. (An informative discussion and visualization can be found in Spivey, 2008, pp. 273–274.) This effect bears some similarity to asymmetry effects in reasoning (i.e., reasoning from a more dominant, or typical, to a less dominant concept; Tversky, 1977).

The present results provide support to a dynamic and thoroughly interactive view of cognition (e.g., Spivey, 2008). In this view, low-level visual representations are under a constant influence of higher-level representations, and this influence leads to activity that is context and task dependent (C. Gilbert & Sigman, 2007; Kveraga, Ghuman, & Bar, 2007; Lamme & Roelfsema, 2000).

The emergent nature of the category effect may also be thought of in terms of retrieving exemplars. Although the stimuli presented to the participants (Bp and Bp) were matched in visual similarity, one can imagine that in categorizing a B, multiple exemplars, including lowercase instances, are activated, rendering the difference between a b and Bwith-activated-exemplars smaller than the difference between a b and Bwithout-activated-exemplars (Cohen & Nosofsky, 2000). Regardless of whether one views the category effect as arising from the activation of multiple specific exemplars, from the activation of a category prototype, or deom neural state-space dynamics that blend prototype and exemplar properties (a view embodied in connectionist accounts, e.g., Hinton & Shallice, 1991), the implications of our results are the same: Performance on an explicitly visual task (physical same/different judgments) is influenced by conceptual categories. This effect depends on processing time, immediately preceding experience, and stimulus typicality. Together, these results constitute evidence for conceptual penetration of visual processing and open the door to examinations of the neural mechanisms by which visual processing is guided by higher-level conceptual representations.

Supplementary Material

Supplementary Materials

Acknowledgments

We thank Emily McDowell, Song-I Yang, and Martekuor Dodoo for their help with data collection.

Funding

This research was supported by a National Science Foundation training grant to G.L. and by National Institutes of Health Grants R01-DC009209 and R01-MH67008 to S.L.T-S and R01-HD049681 to D.S.

Footnotes

Declaration of Conflicting Interests

The authors declared that they had no conflicts of interests with respect to their authorship and/or the publication of this article.

Supplemental Material

Additional supporting information may be found at http://pss.sagepub.com/content/by/supplemental-data

1

We did not include bp trials because they lacked a visual control; pp trials were excluded to ensure a balance of same and different trials. Although this meant that b stimuli were seen more often than p stimuli, this imbalance did not confound the effect because the critical comparison (Bp and Bb) trials were seen equally often. Faster responses to b stimuli due to practice would work against the predicted effect.

References

  1. Cohen A, Nosofsky R. An exemplar-retrieval model of speeded same-different judgments. Journal of Experimental Psychology: Human Perception and Performance. 2000;26:1549–1569. doi: 10.1037//0096-1523.26.5.1549. [DOI] [PubMed] [Google Scholar]
  2. Delorme A, Rousselet G, Mace M, Fabre-Thorpe M. Interaction of top-down and bottom-up processing in the fast visual analysis of natural scenes. Cognitive Brain Research. 2004;19:103–113. doi: 10.1016/j.cogbrainres.2003.11.010. [DOI] [PubMed] [Google Scholar]
  3. Foxe J, Simpson G. Flow of activation from V1 to frontal cortex in humans: A framework for defining "early" visual processing. Experimental Brain Research. 2002;142:139–150. doi: 10.1007/s00221-001-0906-7. [DOI] [PubMed] [Google Scholar]
  4. Gilbert A, Regier T, Kay P, Ivry R. Whorf hypothesis is supported in the right visual field but not the left. Proceedings of the National Academy of Sciences, USA. 2006;103:489–494. doi: 10.1073/pnas.0509868103. [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Gilbert A, Regier T, Kay P, Ivry R. Support for lateralization of the Whorfian effect beyond the realm of color discrimination. Brain and Language. 2008;105:91–98. doi: 10.1016/j.bandl.2007.06.001. [DOI] [PubMed] [Google Scholar]
  6. Gilbert C, Sigman M. Brain states: Top-down influences in sensory processing. Neuron. 2007;54:677–696. doi: 10.1016/j.neuron.2007.05.019. [DOI] [PubMed] [Google Scholar]
  7. Glezer LS, Jiang X, Riesenhuber M. Evidence for highly selective neuronal tuning to whole words in the "visual word form area". Neuron. 2009;62:199–204. doi: 10.1016/j.neuron.2009.03.017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Goldstone R. Influences of categorization on perceptual discrimination. Journal of Experimental Psychology: General. 1994;123:178–200. doi: 10.1037//0096-3445.123.2.178. [DOI] [PubMed] [Google Scholar]
  9. Goldstone R. Perceptual learning. Annual Review of Psychology. 1998;49:585–612. doi: 10.1146/annurev.psych.49.1.585. [DOI] [PubMed] [Google Scholar]
  10. Goldstone R, Steyvers M, Spencer-Smith J, Kersten A. Interactions between perceptual and conceptual learning. In: Diettrich E, Markman A, editors. Cognitive dynamics: Conceptual change in humans and machines. Mahwah, NJ: Erlbaum; 2004. pp. 191–228. [Google Scholar]
  11. Grill-Spector K, Kanwisher N. Visual recognition: As soon as you know it is there, you know what it is. Psychological Science. 2005;16:152–160. doi: 10.1111/j.0956-7976.2005.00796.x. [DOI] [PubMed] [Google Scholar]
  12. Harnad S. Categorical perception: The groundwork of cognition. New York: Cambridge University Press; 1987. [Google Scholar]
  13. Harnad S. Cognition is categorization. In: Cohen H, Lefebvre C, editors. Handbook of categorization. Amsterdam: Elsevier; 2005. pp. 20–45. [Google Scholar]
  14. Hawkins H, Hillyard S, Luck S, Downing C, Mouloua M, Woodward D. Visual-attention modulates signal detectability. Journal of Experimental Psychology: Human Perception and Performance. 1990;16:802–811. doi: 10.1037//0096-1523.16.4.802. [DOI] [PubMed] [Google Scholar]
  15. Hinton GE, Shallice T. Lesioning an attractor network: Investigations of acquired dyslexia. Psychological Review. 1991;98:74–95. doi: 10.1037/0033-295x.98.1.74. [DOI] [PubMed] [Google Scholar]
  16. Hupe J, James A, Girard P, Lomber S, Payne B, Bullier J. Feedback connections act on the early part of the responses in monkey visual cortex. Journal of Neurophysiology. 2001;85:134–145. doi: 10.1152/jn.2001.85.1.134. [DOI] [PubMed] [Google Scholar]
  17. Johnson JS, Olshausen BA. Timecourse of neural signatures of object recognition. Journal of Vision. 2003;3:499–512. doi: 10.1167/3.7.4. Retrieved from http://www.journalofvision.org/3/7/4/ [DOI] [PubMed] [Google Scholar]
  18. Kuhl P. Learning and representation in speech and language. Current Opinions in Neurobiology. 1994;4:812–822. doi: 10.1016/0959-4388(94)90128-7. [DOI] [PubMed] [Google Scholar]
  19. Kveraga K, Ghuman A, Bar M. Top-down predictions in the cognitive brain. Brain and Cognition. 2007;65:145–168. doi: 10.1016/j.bandc.2007.06.007. [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Lamme V, Roelfsema P. The distinct modes of vision offered by feedforward and recurrent processing. Trends in Neurosciences. 2000;23:571–579. doi: 10.1016/s0166-2236(00)01657-x. [DOI] [PubMed] [Google Scholar]
  21. Livingston K, Andrews J, Harnad S. Categorical perception effects induced by category learning. Journal of Experimental Psychology: Learning, Memory, and Cognition. 1998;24:732–753. doi: 10.1037//0278-7393.24.3.732. [DOI] [PubMed] [Google Scholar]
  22. Lupyan G. Reuniting categories, language, and perception. In: McNamara D, Trafton J, editors. Proceedings of the Twenty-Ninth Annual Meeting of the Cognitive Science Society. Austin, TX: Cognitive Science Society; 2007. pp. 1247–1252. [Google Scholar]
  23. Lupyan G. The conceptual grouping effect: Categories matter (and named categories matter more) Cognition. 2008a;108:566–577. doi: 10.1016/j.cognition.2008.03.009. [DOI] [PubMed] [Google Scholar]
  24. Lupyan G. From chair to "chair:" A representational shift account of object labeling effects on memory. Journal of Experimental Psychology: General. 2008b;137:348–369. doi: 10.1037/0096-3445.137.2.348. [DOI] [PubMed] [Google Scholar]
  25. Mack ML, Gauthier I, Sadr J, Palmeri TJ. Object detection and basic-level categorization: Sometimes you know it is there before you know what it is. Psychonomic Bulletin & Review. 2008;15:28–35. doi: 10.3758/pbr.15.1.28. [DOI] [PubMed] [Google Scholar]
  26. McCotter M, Gosselin F, Maccabee PJ, Schyns P. The use of visual information in natural scenes. Visual Cognition. 2005;12:938–953. [Google Scholar]
  27. McMurray B, Aslin RN, Tanenhaus MK, Spivey M, Subik D. Gradient sensitivity to within-category variation in words and syllables. Journal of Experimental Psychology: Human Perception and Performance. 2008;34:1609–1631. doi: 10.1037/a0011747. [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. Mitterer H, Horschig JM, Musseler J, Majid A. The influence of memory on perception: It's not what things look like, it's what you call them. Journal of Experimental Psychology: Learning, Memory, and Cognition. 2009;35:1557–1562. doi: 10.1037/a0017019. [DOI] [PubMed] [Google Scholar]
  29. Norris D, McQueen JM, Cutler A. Merging information in speech recognition: Feedback is never necessary [Target article and commentaries. The Behavioral and Brain Sciences. 2000;23:299–370. doi: 10.1017/s0140525x00003241. [DOI] [PubMed] [Google Scholar]
  30. Notman L, Sowden P, Ozgen E. The nature of learned categorical perception effects: A psychophysical approach. Cognition. 2005;95:B1–B14. doi: 10.1016/j.cognition.2004.07.002. [DOI] [PubMed] [Google Scholar]
  31. Posner M. Chronometric explorations of mind. Hillsdale, NJ: Erlbaum; 1978. [Google Scholar]
  32. Pylyshyn Z. Is vision continuous with cognition? The case for cognitive impenetrability of visual perception. Behavioral and Brain Sciences. 1999;22:341–365. doi: 10.1017/s0140525x99002022. [DOI] [PubMed] [Google Scholar]
  33. Quinn P, Eimas PD, Tarr MJ. Perceptual categorization of cat and dog silhouettes by 3- to 4-month-old infants. Journal of Experimental Child Psychology. 2001;79:78–94. doi: 10.1006/jecp.2000.2609. [DOI] [PubMed] [Google Scholar]
  34. Riesenhuber M, Poggio T. Models of object recognition. Nature Neuroscience. 2000;3:1199–1204. doi: 10.1038/81479. [DOI] [PubMed] [Google Scholar]
  35. Schyns P, Oliva A. Dr. Angry and Mr. Smile: When categorization flexibly modifies the perception of faces in rapid visual presentations. Cognition. 1999;69:243–265. doi: 10.1016/s0010-0277(98)00069-9. [DOI] [PubMed] [Google Scholar]
  36. Schyns P, Rodet L. Categorization creates functional features. Journal of Experimental Psychology: Learning, Memory, and Cognition. 1997;23:681–696. [Google Scholar]
  37. Spivey M. The continuity of mind. Oxford, England: Oxford University Press; 2008. [Google Scholar]
  38. Tversky A. Features of similarity. Psychological Review. 1977;84:327–352. [Google Scholar]
  39. VanRullen R, Thorpe SJ. The time course of visual processing: From early perception to decision-making. Journal of Cognitive Neuroscience. 2001;13:454–461. doi: 10.1162/08989290152001880. [DOI] [PubMed] [Google Scholar]
  40. Winawer J, Witthoft N, Frank M, Wu L, Wade A, Boroditsky L. Russian blues reveal effects of language on color discrimination. Proceedings of the National Academy of Sciences, USA. 2007;104:7780–7785. doi: 10.1073/pnas.0701644104. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplementary Materials

RESOURCES