Abstract
Using functional MRI, we investigated reality monitoring for auditory information. During scanning, healthy young adults heard words in another person’s voice and imagined hearing other words in that same voice. Later, outside the scanner, participants judged words as “heard,” “imagined,” or “new.” An area of left middle frontal gyrus (Brodmann’s area, or BA, 6) was more active at encoding for imagined items subsequently correctly called “imagined” than for items incorrectly called “heard.” An area of left inferior frontal gyrus (BA 45, 44) was more active at encoding for items subsequently called “heard” than “imagined,” regardless of the actual source of the item. Scores on an Auditory Hallucination Experience Scale were positively related to activity in superior temporal gyrus (BA 22) for imagined words incorrectly called “heard.” We suggest that activity in these areas reflects cognitive operations information (middle frontal gyrus) and semantic and perceptual detail (inferior frontal gyrus and superior temporal gyrus, respectively) used to make reality-monitoring attributions.
Keywords: reality monitoring, source monitoring, false memory, imagination, auditory hallucination
Source monitoring refers to mechanisms by which people make attributions about the origin of mental experiences. According to the source-monitoring framework (SMF; Johnson, Hashtroudi, & Lindsay, 1993), this includes evaluating qualitative characteristics such as perceptual, semantic, temporal, and spatial detail and information about the cognitive operations engaged during encoding. In the present study, we were concerned with one class of such attributions, reality monitoring—discriminating perceived from self-generated events (e.g., differentiating a memory of a previous perception from that of a previous imagination; Johnson & Raye, 1981).
Reality monitoring is an essential but imperfect cognitive function (Johnson & Raye, 1981). External misattributions of self-generated information (i.e., reality-monitoring errors) are fairly common, and much has been learned from systematically studying such errors in the laboratory (see Johnson, 2006; Johnson, Raye, Mitchell, & Ankudowich, 2011; Lindsay, 2008, for reviews). For example, good imagers make more misattributions of self-generated information to an external source than do poor imagers, presumably because good imagers create representations that are rich in visual and other contextual details, which makes them similar to perceptions, or they do so easily or spontaneously, which leaves few records of cognitive effort (Dobson & Markham, 1993; Johnson, Raye, Wang, & Taylor, 1979). Records of the cognitive operations (e.g., the act of imagining) carried out at the time of encoding can be used during remembering as a cue that the information was self-generated. Anything that decreases production of, or attention to, cognitive operations information is likely to increase errors (Finke, Johnson, & Shyi, 1988). Hence, it is likely that both more vivid perceptual information and less vivid information about cognitive operations play a role in misattributing self-generated information to perception.
The idea that reality-monitoring errors are based, at least in part, on self-generated perceptual information is supported by recent functional MRI (fMRI) studies. Gonsalves et al. (2004) scanned participants as they saw pictures and imagined pictures. A subsequent memory analysis identified an area of precuneus—a brain region associated with visual processing—where activity was greater for imagined items that participants subsequently incorrectly called “seen” than for those items correctly called “imagined” (see also, e.g., Kensinger & Schacter, 2005). In addition, activity in posterior visual areas during test is associated with attributions to perception, regardless of the actual source (i.e., seen vs. imagined pictures; Kensinger & Schacter, 2006). In contrast, activity in left lateral and medial prefrontal cortex (PFC) at both encoding (e.g., Kensinger & Schacter, 2005) and test (e.g., Dobbins & Wagner, 2005; Kensinger & Schacter, 2006; Mitchell et al., 2008; Simons, Gilbert, Owen, Fletcher, & Burgess, 2005) is associated with correct attributions of self-generated information. Presumably, these PFC areas process or represent either the cognitive operations engaged during encoding (e.g., those involved in imagining or in carrying out a particular encoding task, such as estimating cost) or the conceptual-semantic output of these processes.
Most studies of reality monitoring in healthy individuals have used visual information, but reality-monitoring errors occur for auditory information as well (Foley, Johnson, & Raye, 1983; Johnson, Foley, & Leach, 1988). For example, Johnson et al. (1988) found that participants were more likely to misattribute imagined words to a speaker if they had imagined the words in the speaker’s voice than if they had imagined them in their own voice. Evidence from fMRI regarding the specific mechanisms of misattributions of auditory information in healthy individuals, however, is missing. Hence, the primary aim of the current study was to specify brain mechanisms underlying external misattributions of auditory information in healthy individuals.
In the scanner, healthy young adults heard some words in another person’s voice and imagined hearing other words in that person’s voice. We expected that increased activity in brain areas associated with hearing (primary and secondary auditory cortex) should be associated with heard items, and increased activity in areas associated with cognitive operations or the output of those operations such as semantic information (e.g., medial and left lateral prefrontal cortex) should be associated with imagined items. Later, we presented participants with visual words and asked them whether they heard each word, they imagined it, or it was new: Our primary interest was in identifying which of the areas with differential encoding activity predicted subsequent reality-monitoring performance. We expected that activity in areas associated with cognitive operations would be related to correct attributions of imagined items. We also expected, parallel to the findings of Gonsalves et al. (2004), that misattributions of imagined words as heard would be related to activity in areas associated with representing the output of such imagining (auditory or semantic detail).
Although most studies have investigated reality monitoring in the domain of memory, similar factors and mechanisms likely contribute to reality monitoring in the domains of perception and belief (Johnson, 1988). That is, the SMF is relevant for understanding false perceptions and beliefs (e.g., hallucinations and delusions; see Woodward & Menon, 2013, for a review and discussion) as well as false memories (Johnson et al., 2011). Consistent with this idea, studies have found that nonclinical participants who are prone to auditory hallucinations show increased external misattributions of self-generated information in various reality-monitoring tasks, relative to participants who are not prone to hallucinations (e.g., Larøi, Van der Linden, & Marczewski, 2004; Sugimori, Asai, & Tanno, 2011). Thus, we also assessed the relation between participants’ proneness to auditory hallucinations and brain activity associated with misattributions of imagined auditory information. This should help further refine the understanding of the overlap in mechanisms underlying reality-monitoring errors and online misperceptions (e.g., poor cognitive operations information, overly vivid auditory information). Given evidence from the clinical literature about the importance of superior temporal gyrus (STG) in auditory hallucinations in patients (Kompus, Westerhausen, & Hugdahl, 2011), we expected a relation between proneness to auditory hallucinations and STG activity.
Method
Participants
Twenty participants (11 females, 9 males; all right-handed; mean age = 22.2 years, SD = 3.8, range = 18–29) self-reported being in good health, with no history of stroke, heart disease, primary degenerative neurological disorder, or psychiatric diagnosis; they had normal or corrected-to-normal vision and were not taking psychotropic medications. All participants were paid. The Human Investigation Committee of Yale University Medical School approved the protocol; informed consent was obtained from all participants.
Design
There were two phases in the study. Phase 1 (encoding) was carried out while participants were scanned. There were three types of trials: heard words, imagined words, and shapes. Phase 2 (test) was a surprise source-memory test conducted outside the scanner; participants responded “heard,” “imagined,” or “new” to heard, imagined, and new items, respectively.
Materials
A total of 308 words were used: 10 for practice outside the scanner, 10 for practice inside the scanner, and 288 for experimental trials. The experimental words were assigned to three 96-word lists equated on several characteristics, such as frequency, concreteness, familiarity, imageability, meaningfulness, and number of letters, phonemes, and syllables, using the MRC Psycholinguistic Database (Coltheart, 1981). The three lists were rotated across the heard, imagined, and new conditions. A recorded male voice read aloud a news story, instructions for Phase 1, and the words for the heard condition. During encoding, we also presented 30 nonauditory shape trials: figures composed of four contiguous shapes (with various numbers of circular and square shapes).
We used an Auditory Hallucination Experience Scale (AHES) to measure auditory verbal-hallucination proneness. It had 12 items adapted from the Launay Slade Hallucination Scale (Launay & Slade, 1981), a widely used instrument for measuring disposition to hallucinate. One statement measuring predisposition to visual hallucination (“On occasion, I have seen a person’s face in front of me when no one was in fact there”) was deleted; “I have heard the voice of the Devil” and “In the past, I have heard the voice of God speaking to me” were replaced with “I have heard a voice inside me calling me toward bad things” and “I have heard a voice inside me telling me to do good things,” respectively.
Procedure
Task instructions, read in the same voice as the words in the heard condition, were given outside the scanner via headphones. Participants were told this was a study comparing imagination with perception. To ensure participants were familiar with the sound of the voice they would later be asked to imagine, we also played them a recording of the voice reading a news story for approximately 3.5 min.
On each trial in Phase 1, participants first saw a cue in the middle of the computer screen for 2 s that told them whether they would hear a word (heard condition), they should imagine hearing a word in the pre-exposed “other” voice (imagined condition), or they would see a shape (shape condition). The cues “hear” and “imagine” were each followed by a new word presented visually for 2 s, and concurrently, participants either heard the male voice saying the word or imagined his voice saying that word. When the word disappeared from the screen, participants were cued to rate, by pressing buttons with the first three fingers of their right hand, how well they heard or imagined hearing the word (1 = low quality, 2 = average, 3 = high quality). Participants had 2 s to make their response.
The “shape” cue was followed by a figure composed of four adjoined circular and square shapes presented visually for 2 s. When the figure disappeared from the screen, participants were cued to estimate, using the same three buttons, whether there were more square or circular shapes (1 = more square shapes, 2 = an equal number of square and circular shapes, 3 = more circular shapes). Participants had 2 s to make their response.
Between trials, there was an unfilled 6-s intertrial interval to allow the hemodynamic response to return to baseline. Participants practiced five heard trials, five imagined trials, and three shape trials in pseudorandom order before getting into the scanner and practiced the same number of trials again (with different items) during the structural scans so that the volume of the headphones could be adjusted as needed for each participant. Phase 1 consisted of six runs in the scanner. In each run, participants heard 16 words, imagined the voice saying 16 words, and saw 5 shapes; these trials were pseudorandomly intermixed.
About 5 min after the scan, there was a surprise source-memory test in a separate room. This test consisted of 96 heard items, 96 imagined items, and 96 new items (which had not been presented elsewhere in the experiment) sequentially presented visually on a computer screen for 4 s each in pseudorandom order. Participants indicated, for each word, if they had heard it, if they had imagined it, or if it was new, pressing the keys 1, 2, and 3, respectively, using the first three fingers of their right hand.
Imaging details
Images were acquired using a 3T Siemens Trio scanner. After anatomical localizer scans, functional images were acquired with a single-shot echoplanar gradient-echo pulse sequence (repetition time = 2,000 ms, echo time = 25 ms, flip angle = 80°, field of view = 240 mm). The 36 oblique axial slices were 3.5 mm thick with an in-plane resolution of 3.75 × 3.75 mm; they were aligned with the anterior commissure-posterior commissure line. Each run began with 12 s of blank screen to allow tissue to reach steady-state magnetization and was followed by a 1-min rest interval.
fMRI analyses
After reconstruction, time series were shifted by sinc interpolation to correct for slice-acquisition times. Data were motion-corrected using a 6-parameter automated algorithm (Automated Image Registration, or AIR; Woods, Cherry, & Mazziotta, 1992). A 12-parameter AIR algorithm was used to coregister participants’ images to a common reference brain. Data were mean-normalized across time and participant and spatially smoothed (3-D, 8-mm, full-width half-maximum Gaussian kernel). We used voxel-based analysis of variance (ANOVA) with participant as a random factor and all other factors fixed (NeuroImaging Software; Laboratory for Clinical Cognitive Neuroscience, University of Pittsburgh, and Center for the Study of Brain, Mind, and Behavior, Princeton University). F maps were transformed to Talairach space using Analysis of Functional NeuroImages (AFNI) software (http://afni.nimh.nih.gov/afni), and areas of activation were localized using AFNI and Talairach Daemon software (Lancaster et al., 1997, 2000) and were visually checked against atlases.
Because we were interested in event-related responses, we first identified regions showing an Encoding Condition (heard words, imagined words, shapes)1 × Time Within Trial (Scans 1–6) interaction with a minimum of 10 contiguous voxels, each significant at p < 1.0 × 10−13 (Lieberman & Cunningham, 2009). For the regions thus identified, subsequent memory analyses were conducted on the mean percentage change (from Time 1) at Times 3 and 4 or Times 4 and 5 (depending on maximal differences in activity per region), using a 2 (encoding condition: heard words, imagined words) × 3 (response at test: “heard,” “imagined,” “new”) ANOVA. We also calculated, for these same regions, Pearson’s correlation coefficients between these mean percentage change scores for each participant for each combination of encoding condition and subsequent response, and scores on the AHES.
Results
Behavioral data
On-line ratings (maximum = 3.00) indicated that participants were able to hear and imagine the words well. Participants gave higher ratings to heard items (M = 2.71) than to imagined items (M = 2.50), F(1, 19) = 15.57, p = .001. They also gave higher ratings to items they subsequently called “heard” (M = 2.68) than to those they called “imagined” (M = 2.59) or “new” (M = 2.56), F(2, 38) = 7.76, p = .001; the ratings of items called “imagined” did not differ significantly from those called “new.”
Figure 1 shows mean proportion of each response type during the source-memory test for heard, imagined, and new items. For old words misattributed as new, there was no significant difference between the proportion of items that had been encoded in the heard condition (M = 0.43) versus the imagined condition (M = 0.47), F(1, 19) = 0.36, p = .55, ηp2 = .019. There was also no significant difference in the proportion of new words misattributed as heard (M = 0.14) versus imagined (M = 0.13), F(1, 19) = 0.11, p = .74, ηp2 = .006. Thus old/new recognition was approximately the same for heard and imagined items.
Of central interest, for old items, a 2 (item type: heard words, imagined words) × 2 (response: “heard,” “imagined”) ANOVA was conducted. There was a significant interaction between item type and response, F(1, 19) = 7.33, p = .01, ηp2 = .278. For heard items, participants made a higher proportion of correct “heard” responses (M = 0.34) than incorrect “imagined” responses (M = 0.23), F(1, 38) = 5.84, p = .02, ηp2 = .133. For imagined items, there was no significant difference between the proportion of correct “imagined” responses (M = 0.28) and incorrect “heard” responses (M = 0.26; p = .60).2 This pattern is remarkably similar to the pattern seen in Johnson et al. (1988, Experiment 2), in which old/new recognition was much better (hits ~75% did not differ across the heard and imagined conditions), which indicates that reality-monitoring difficulty does not depend on low levels of recognition.
From the behavioral data, it might be tempting to conclude that participants had some source information about heard words but none about imagined words and hence simply guessed the source on the imagined items recognized as old. However, the fMRI data provide evidence that participants were not simply guessing on the imagined items but rather, as predicted by the SMF, were responding based on qualities of their subjective experience.
fMRI data
Table 1 shows all areas identified in our initial whole-brain analysis. The section headings in the table indicate the main effects of encoding condition identified by the subsequent memory analyses. Of primary interest were areas that showed a significant main effect of response type at test or an Encoding Condition × Response Type interaction.
Table 1.
Hemisphere | Brodmann’s area | Anatomical area | Talairach coordinates | Maximum F for the region | Effect of encoding condition in subsequent analyses | ||||
---|---|---|---|---|---|---|---|---|---|
|
|
|
|||||||
x | y | z | F(10, 190) | ηp2 | F(1, 19)a | ηp2 | |||
Imagined words > heard words | |||||||||
Left | 45, 44 | Inferior frontal gyrus | −47 | 15 | 19 | 24.16 | .560 | 21.96 | .536 |
Left | 6, (4) | Middle frontal gyrus (precentral gyrus) | −29 | −10 | 48 | 13.67 | .418 | 15.81 | .454 |
Right | 6 | Middle frontal gyrus (superior frontal gyrus) | 23 | −5 | 50 | 18.87 | .498 | 6.55 | .256 |
Right | 6 | Precentral gyrus | 46 | 2 | 28 | 19.28 | .503 | 4.76 | .200 |
Medial | 31, 19 | Precuneus, cingulate gyrus | −8 | −58 | 29 | 15.71 | .453 | 2.99 | .136 |
Left | 19, 18, 7 | Middle occipital gyrus, superior occipital gyrus, fusiform gyrus, inferior parietal lobule | −31 | −73 | 24 | 39.23 | .674 | 6.53 | .256 |
Right | 7, 19, 37, 39, 40 | Precuneus, middle occipital gyrus, superior occipital gyrus, fusiform gyrus, supramarginal gyrus, angular gyrus, inferior parietal lobule, superior parietal lobule | 21 | −66 | 38 | 67.33 | .780 | 8.79 | .316 |
| |||||||||
Heard words > imagined words | |||||||||
Left | 22, 21, 39, 37 | Superior temporal gyrus, middle temporal gyrus, tranverse temporal gyrus, inferior parietal lobule | −46 | −22 | 6 | 64.33 | .772 | 139.05 | .880 |
Right | 22, 21, 39, 37 | Superior temporal gyrus, middle temporal gyrus, tranverse temporal gyrus, inferior parietal lobule | 46 | −23 | 57 | 71.63 | .790 | 145.45 | .884 |
Note: Bolded areas are discussed in the text. Talairach coordinates are shown for the voxel with the maximum F value in each area of activation. For identified areas, subsequent memory analyses were conducted on the mean percentage change (from Time 1) at Times 3 through 4 or Times 4 through 5 (depending on maximal differences in activity), using a 2 (encoding condition: heard words, imagined words) × 3 (response type at test: “heard,” “imagined,” “new”) analysis of variance. The main effects of encoding condition for these areas are indicated in the table’s section headings, along with the F values in the rightmost column. Main effects of response type at test and interactions are reported in the text. For each area of activation, the major anatomical regions and Brodmann’s area numbers are listed starting with the location of the local maxima and extending from there (parentheses indicate a small extent relative to other areas listed).
ps < .05.
An area including left middle frontal gyrus (MFG), extending into precentral gyrus—Brodmann’s area (BA) 6, (4) showed a significant interaction between encoding condition and response type at Times 3 through 4, F(2, 38) = 3.36, p = .04, ηp2 = .150 (Fig. 2a). For heard items, there was no significant difference in activity among responses (F < 1), but for imagined items, F(2, 38) = 4.05, p = .03, ηp2 = .176, there was more activity for items subsequently correctly called “imagined” than for items incorrectly called “heard,” F(1, 19) = 4.88, p = .03, ηp2 = .204, or “new,” F(1, 19) = 7.06, p = .02, ηp2 = .271; “heard” and “new” responses did not differ significantly (p > .05).
An area including left inferior frontal gyrus (IFG; BA 45, 44; Fig. 2b) showed a significant main effect of response type at Times 4 through 5, F(2, 38) = 3.62, p = .04, ηp2 = .167: “Heard” responses were greater than “imagined” responses, F(1, 19) = 6.34, p = .04, ηp2 = .251, or “new” responses, F(1, 19) = 4.32, p = .02, ηp2 = .185; “imagined” and “new” responses were not significantly different (p > .05). The Encoding Condition × Response Type interaction was not significant (Fs < 1).
We were also interested in areas that showed correlations with the AHES. There were only two such areas: Large bilateral areas of STG that extended to include middle and transverse temporal gyri and inferior parietal lobule (Figs. 3a and 3b) showed only main effects of encoding condition at Times 4 through 5 (heard > imagined; see Table 1). No other subsequent memory effects were significant. As shown in Figures 3c and 3d, activity in these regions for imagined items called “heard” was positively related to scores on the AHES (right STG: r = .54, p = .01; left STG: r = .46, p = .04). People more prone to auditory hallucinations showed more activation in this region for imagined items later mistakenly called “heard.”
Discussion
Consistent with expectations that frontal regions would reflect engaging cognitive operations, our results showed that at encoding, there was greater activity for imagined than heard words in left MFG and left IFG. Consistent with the expectation that temporal regions would reflect processing auditory information, areas including bilateral STG were more active for heard than imagined words. Our main question was how brain activity in these regions during encoding was related to subsequent reality-monitoring judgments for whether words were heard or imagined.
Greater activity in left MFG (BA 6, [4]) during encoding for imagined items was associated with later correctly calling the items “imagined” rather than incorrectly calling them “heard.” This pattern suggests that records of encoding activity in this area as participants imagined the words functioned later as cues at test. Similar areas are frequently seen in working memory tasks, especially those that require manipulation of information (e.g., Hanakawa et al., 2002; see Cabeza & Nyberg, 2000, for a review), which suggests that this area is associated with engaging cognitive operations. Hence, the pattern is consistent with the prediction from the SMF (Johnson et al., 1993) that records of cognitive operations carried out at the time of encoding (in this case, imagining) can later be used to judge an event as self-generated rather than perceived.3
Another (perhaps more speculative) possibility consistent with the SMF is that, in this paradigm, activity in this region of MFG represents cognitive operations specifically involved in localizing active auditory representations or the representation of location information. In effect, reality monitoring of speech requires “location” cues that differentiate internal from external sources. Further, it has been suggested that hallucinations reflect difficulty with the “subjective spatial location” of imagined events (Woodward & Menon, 2013, p. 171). There is evidence for a dorsal “where” auditory PFC pathway that is active during sound localization in both perceptual and working memory tasks (see Arnott & Alain, 2011, for a review). For example, an area with a local maximum about 1 voxel from our BA 6 region showed increased activity in an n-back task requiring participants to respond on the basis of sound location rather than sound identity (Leung & Alain, 2011). The current BA 6 is also similar to an area found during the delay period of a working memory task when participants had to hold auditory locations in mind versus a saccade localizer (Tark & Curtis, 2009) and when participants did a sound location delayed-match-to-sample task versus a sound identity-match-to-sample task (Arnott, Grady, Hevenor, Graham, & Alain, 2005).
The fact that left IFG (BA 45, 44) was more active at encoding for imagined than for heard words is consistent with the association of this area with speech generation (e.g., Horwitz et al., 2003); thus, one might expect the records of the cognitive operations associated with generating the auditory images to provide cues that items were imagined. However, encoding activity in this area was greater for items later called “heard” than “imagined” regardless of the source of the item; records of the activity in this region obviously were not serving as a cue that information was self-generated. As noted in the introduction, IFG may represent information generated (e.g., semantic information) as well as (or, in some subregions, instead of) the mode by which it is derived (see Mitchell et al., 2008). In fact, left IFG is associated not only with language generation but also with language comprehension (Skipper, Goldin-Meadow, Nusbaum, & Small, 2007) and semantic processing (see Gabrieli, Poldrack, & Desmond, 1998, for a review). Furthermore, encoding activity in a similar area was associated with subsequent true and false memory in a study using a procedure designed to induce false memories for semantic associates of presented words (Kim & Cabeza, 2007). According to the SMF, greater semantic information is usually associated with perceived than with imagined events (Johnson & Raye, 1981). Thus, it is possible that source judgments at test were primarily sensitive to the amount of semantic information activated and not the amount of cognitive operations information. Clearly, the conditions determining the relative roles of IFG in cognitive operations as opposed to representation of semantic information during source monitoring remain to be specified as the functional and structural heterogeneity of IFG is clarified (cf. Fedorenko, Duncan, & Kanwisher, 2012).
In any event, the pattern of brain activity highlights the added value of neuroimaging data. Although, for imagined items, participants’ behavioral “imagined” and “heard” responses did not differ, “heard” responses were related to IFG activity at encoding and “imagined” responses to MFG activity. These are areas one would expect to be associated with the processing of semantic and cognitive operations information, respectively. Hence, even with no difference in “heard” and “imagined” judgments, the brain data suggest that, as predicted by the SMF, participants were likely basing these reality-monitoring attributions on the relative value of different types of information.
Support for this proposition comes also from the AHES correlations with STG activity (an auditory area) only for imagined items called “heard.” Bilateral STG (extending to adjacent areas) activated when participants heard words and, on average, it deactivated when they imagined hearing words in another person’s voice. Assuming that STG represents auditory information, the SMF would predict that differences in activity for heard and imagined items should provide a basis for reality-monitoring judgments. Thus, increased activity in this region during imagination would decrease the availability of an important cue about the origin of auditory information. If so, activity in STG during auditory imagination may provide a useful marker of proneness to hallucinations. Consistent with this idea, one of our most interesting findings was that scores on the AHES were positively related to activity in these areas for imagined items that participants subsequently incorrectly called “heard.” It may be that people prone to auditory hallucinations are good auditory imagers—that is, they relatively effortlessly or spontaneously produce vivid auditory imaginations that rival those of actually heard words. In clinical populations, individuals with schizophrenia who experience auditory verbal hallucinations are more likely to confuse what they have heard and what they imagined hearing in a laboratory reality-monitoring task than those without a history of auditory verbal hallucinations (Brunelin et al., 2006). Moreover, STG activity is associated with auditory hallucinations in patients (e.g., Bentaleb, Beauregard, Liddle, & Stip, 2002; Dierks et al., 1999; van de Ven et al., 2005), and hyperactivity of STG, perhaps due to reduced control from PFC, has been at the center of several theories of auditory hallucinations (see Woodward & Menon, 2013, for a recent review and discussion). Consistent with predictions based on the SMF that differences in activity in perceptual-processing areas should provide a basis for reality-monitoring judgments, a recent meta-analysis (Kompus et al., 2011) showed that patients with auditory hallucinations demonstrate both decreased activation in STG during external (i.e., perceptual) auditory stimulation and increased activation in the absence of external auditory stimulation.
In summary, the current findings provide new evidence regarding the processes affecting auditory reality monitoring and source misattribution of auditory information. We found that brain activity during encoding of words predicted whether people later mistakenly responded “heard” when presented with imagined items. Consistent with the SMF, results showed that less activity in a brain area associated with cognitive operations (left MFG) and more activity in a brain area associated with semantic processing (left IFG) during imagining another person’s voice led healthy people to later judge imagined items as “heard.” Further, the more participants were prone to have auditory-hallucination-like experiences, the more brain areas associated with speech perception (right and left STG, BA 22) were active for imagined items subsequently called “heard.” The findings thus contribute to the understanding of the neural correlates of reality-monitoring errors for auditory information.
Acknowledgments
We thank the magnetic-resonance technicians at the Yale Magnetic Resonance Research Center for help in data collection, and Matt Sazma for help with figures.
Funding
This research was funded by the National Institute on Aging (Grant R37AG009253) to M. K. Johnson.
Footnotes
The shape condition was included in the initial ANOVA so that areas related to both heard and imagined words might be identified; however, there were no areas where activations for heard and imagined words were equal. The shape condition was of no further interest, and those data were not included in the subsequent memory analyses.
There also was no difference between “heard” and “imagined” responses for new items, which argues against an overall bias to call items “heard” whenever participants were not sure about the source.
We did not find differences in the medial PFC area associated with recollection of cognitive operations in other reality-monitoring studies (e.g., Turner, Simons, Gilbert, Frith, & Burgess, 2008; Vinogradov et al., 2006). One possible reason may be related to the auditory imagination requirements of our task versus other kinds of generation (e.g., solving anagrams) used in other studies. Another possibility is that medial PFC may come online during monitoring at test, not during initial generation of cognitive operations. Studies aimed at reconciling these differences should be informative.
Author Contributions
E. Sugimori developed the study concept. E. Sugimori, M. K. Johnson, K. J. Mitchell, and C. L. Raye contributed to the study design. E. Sugimori and E. J. Greene prepared the materials, and E. Sugimori programmed the experiment, with input from M. K. Johnson, K. J. Mitchell, and C. L. Raye. Data were collected by E. Sugimori and K. J. Mitchell. E. J. Greene performed the functional MRI analyses, and E. Sugimori performed the behavioral data analyses under the supervision of M. K. Johnson and K. J. Mitchell. E. Sugimori, M. K. Johnson, K. J. Mitchell, and C. L. Raye contributed to the interpretation and the writing of the manuscript. All authors approved the final version of the manuscript for submission.
Declaration of Conflicting Interests
The authors declared that they had no conflicts of interest with respect to their authorship or the publication of this article.
References
- Arnott SR, Alain C. The auditory dorsal pathway: Orienting vision. Neuroscience and Biobehavioral Reviews. 2011;35:2162–2173. doi: 10.1016/j.neubiorev.2011.04.005. [DOI] [PubMed] [Google Scholar]
- Arnott SR, Grady CL, Hevenor SJ, Graham S, Alain C. The functional organization of auditory working memory as revealed by fMRI. Journal of Cognitive Neuroscience. 2005;17:819–831. doi: 10.1162/0898929053747612. [DOI] [PubMed] [Google Scholar]
- Bentaleb LA, Beauregard M, Liddle P, Stip E. Cerebral activity associated with auditory verbal hallucinations: A functional magnetic resonance imaging case study. Journal of Psychiatry and Neuroscience. 2002;27:110–115. [PMC free article] [PubMed] [Google Scholar]
- Brunelin J, Combris M, Poulet E, Kallel L, D’Amato T, Dalery J, Saoud M. Source monitoring deficits in hallucinating compared to non-hallucinating patients with schizophrenia. European Psychiatry. 2006;21:259–261. doi: 10.1016/j.eurpsy.2006.01.015. [DOI] [PubMed] [Google Scholar]
- Cabeza R, Nyberg L. Imaging cognition II: An empirical review of 275 PET and fMRI studies. Journal of Cognitive Neuroscience. 2000;12:1–47. doi: 10.1162/08989290051137585. [DOI] [PubMed] [Google Scholar]
- Coltheart M. The MRC Psycholinguistic Database. The Quarterly Journal of Experimental Psychology A. 1981;33:497–505. doi: 10.1080/14640748108400805. [DOI] [Google Scholar]
- Dierks T, Linden DE, Jandl M, Formisano E, Goebel R, Lanfermann H, Singer W. Activation of Heschl’s gyrus during auditory hallucinations. Neuron. 1999;22:615–621. doi: 10.1016/s0896-6273(00)80715-1. [DOI] [PubMed] [Google Scholar]
- Dobbins IG, Wagner AD. Domain-general and domain-sensitive prefrontal mechanisms for recollecting events and detecting novelty. Cerebral Cortex. 2005;15:1768–1778. doi: 10.1093/cercor/bhi054. [DOI] [PubMed] [Google Scholar]
- Dobson M, Markham R. Imagery ability and source monitoring: Implications for eyewitness memory. British Journal of Psychology. 1993;84:111–118. doi: 10.1111/j.2044-8295.1993.tb02466.x. [DOI] [PubMed] [Google Scholar]
- Fedorenko E, Duncan J, Kanwisher N. Language-selective and domain-general regions lie side by side within Broca’s area. Current Biology. 2012;22:2059–2062. doi: 10.1016/j.cub.2012.09.011. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Finke RA, Johnson MK, Shyi GC. Memory confusions for real and imagined completions of symmetrical visual patterns. Memory & Cognition. 1988;16:133–137. doi: 10.3758/bf03213481. [DOI] [PubMed] [Google Scholar]
- Foley MA, Johnson MK, Raye CL. Age-related changes in confusion between memories for thoughts and memories for speech. Child Development. 1983;54:51–60. [PubMed] [Google Scholar]
- Gabrieli JD, Poldrack RA, Desmond JE. The role of left prefrontal cortex in language and memory. Proceedings of the National Academy of Sciences, USA. 1998;95:906–913. doi: 10.1073/pnas.95.3.906. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gonsalves B, Reber PJ, Gitelman DR, Parrish TB, Mesulam MM, Paller KA. Neural evidence that vivid imagining can lead to false remembering. Psychological Science. 2004;15:655–660. doi: 10.1111/j.0956-7976.2004.00736.x. [DOI] [PubMed] [Google Scholar]
- Hanakawa T, Honda M, Sawamoto N, Okada T, Yonekura Y, Fukuyama H, Shibasaki H. The role of rostral Brodmann area 6 in mental-operation tasks: An integrative neuroimaging approach. Cerebral Cortex. 2002;12:1157–1170. doi: 10.1093/cercor/12.11.1157. [DOI] [PubMed] [Google Scholar]
- Horwitz B, Amunts K, Bhattacharyya R, Patkin D, Jeffries K, Zilles K, Braun AR. Activation of Broca’s area during the production of spoken and signed language: A combined cytoarchitectonic mapping and PET analysis. Neuropsychologia. 2003;41:1868–1876. doi: 10.1016/s0028-3932(03)00125-8. [DOI] [PubMed] [Google Scholar]
- Johnson MK. Discriminating the origin of information. In: Oltmanns TF, Maher BA, editors. Delusional beliefs. New York, NY: Wiley & Sons; 1988. pp. 34–65. [Google Scholar]
- Johnson MK. Memory and reality. American Psychologist. 2006;61:760–771. doi: 10.1037/0003-066X.61.8.760. [DOI] [PubMed] [Google Scholar]
- Johnson MK, Foley MA, Leach K. The consequences for memory of imagining in another person’s voice. Memory & Cognition. 1988;16:337–342. doi: 10.3758/bf03197044. [DOI] [PubMed] [Google Scholar]
- Johnson MK, Hashtroudi S, Lindsay DS. Source monitoring. Psychological Bulletin. 1993;114:3–28. doi: 10.1037/0033-2909.114.1.3. [DOI] [PubMed] [Google Scholar]
- Johnson MK, Raye CL. Reality monitoring. Psychological Review. 1981;88:67–85. [Google Scholar]
- Johnson MK, Raye CL, Mitchell KJ, Ankudowich E. The cognitive neuroscience of true and false memories. In: Belli RF, editor. Nebraska Symposium on Motivation: Vol. 58. True and false recovered memories: Toward a reconciliation of the debate. New York, NY: Springer; 2011. pp. 15–52. [DOI] [PubMed] [Google Scholar]
- Johnson MK, Raye CL, Wang AY, Taylor TH. Fact and fantasy: The roles of accuracy and variability in confusing imaginations with perceptual experiences. Journal of Experimental Psychology: Human Learning and Memory. 1979;5:229–240. [PubMed] [Google Scholar]
- Kensinger EA, Schacter DL. Emotional content and reality-monitoring ability: fMRI evidence for the influences of encoding processes. Neuropsychologia. 2005;43:1429–1443. doi: 10.1016/j.neuropsychologia.2005.01.004. [DOI] [PubMed] [Google Scholar]
- Kensinger EA, Schacter DL. Neural processes underlying memory attribution on a reality-monitoring task. Cerebral Cortex. 2006;16:1126–1133. doi: 10.1093/cercor/bhj054. [DOI] [PubMed] [Google Scholar]
- Kim H, Cabeza R. Differential contributions of prefrontal, medial temporal, and sensory-perceptual regions to true and false memory formation. Cerebral Cortex. 2007;17:2143–2150. doi: 10.1093/cercor/bhl122. [DOI] [PubMed] [Google Scholar]
- Kompus K, Westerhausen R, Hugdahl K. The “paradoxical” engagement of the primary auditory cortex in patients with auditory verbal hallucinations: A meta-analysis of functional neuroimaging studies. Neuropsychologia. 2011;49:3361–3369. doi: 10.1016/j.neuropsychologia.2011.08.010. [DOI] [PubMed] [Google Scholar]
- Lancaster JL, Rainey LH, Summerlin JL, Freitas CS, Fox PT, Evans AC, Toga AW, Mazziotta JC. Automated labeling of the human brain: A preliminary report on the development and evaluation of a forward-transform method. Human Brain Mapping. 1997;5:238–242. doi: 10.1002/(SICI)1097-0193(1997)5:4<238::AID-HBM6>3.0.CO;2-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lancaster JL, Woldorff MG, Parsons LM, Liotti M, Freitas CS, Rainey L, Kochunov PV, Nickerson D, Mikiten SA, Fox PT. Automated Talairach Atlas labels for functional brain mapping. Human Brain Mapping. 2000;10:120–131. doi: 10.1002/1097-0193(200007)10:3<120::AID-HBM30>3.0.CO;2-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Larøi F, Van der Linden M, Marczewski P. The effects of emotional salience, cognitive effort and meta-cognitive beliefs on a reality monitoring task in hallucination-prone subjects. British Journal of Clinical Psychology. 2004;43:221–233. doi: 10.1348/0144665031752970. [DOI] [PubMed] [Google Scholar]
- Launay G, Slade P. The measurement of hallucinatory predisposition in male and female prisoners. Personality and Individual Differences. 1981;2:221–234. [Google Scholar]
- Leung AWS, Alain C. Working memory load modulates the auditory “what” and “where” neural networks. NeuroImage. 2011;55:1260–1269. doi: 10.1016/j.neuroimage.2010.12.055. [DOI] [PubMed] [Google Scholar]
- Lieberman MD, Cunningham WA. Type I and Type II error concerns in fMRI research: Re-balancing the scale. Social Cognitive and Affective Neuroscience. 2009;4:423–428. doi: 10.1093/scan/nsp052. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lindsay DS. Source monitoring. In: Byrne JH, Roediger HL III, editors. Learning and memory: A comprehensive reference. Vol. 2. Oxford, England: Elsevier; 2008. pp. 325–347. [Google Scholar]
- Mitchell KJ, Raye CL, McGuire JT, Frankel H, Greene EJ, Johnson MK. Neuroimaging evidence for agenda-dependent monitoring of different features during short-term source memory tests. Journal of Experimental Psychology: Learning, Memory, and Cognition. 2008;34:780–790. doi: 10.1037/0278-7393.34.4.780. [DOI] [PubMed] [Google Scholar]
- Simons JS, Gilbert SJ, Owen AM, Fletcher PC, Burgess PW. Distinct roles for lateral and medial anterior prefrontal cortex in contextual recollection. Journal of Neurophysiology. 2005;94:813–820. doi: 10.1152/jn.01200.2004. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Skipper JI, Goldin-Meadow S, Nusbaum HC, Small SL. Speech-associated gestures, Broca’s area, and the human mirror system. Brain and Language. 2007;101:260–277. doi: 10.1016/j.bandl.2007.02.008. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Sugimori E, Asai T, Tanno Y. Sense of agency over speech and proneness to auditory hallucinations: The reality monitoring paradigm. The Quarterly Journal of Experimental Psychology. 2011;64:169–185. doi: 10.1080/17470218.2010.489261. [DOI] [PubMed] [Google Scholar]
- Tark KJ, Curtis CE. Persistent neural activity in the human frontal cortex when maintaining space that is off the map. Nature Neuroscience. 2009;12:1463–1468. doi: 10.1038/nn.2406. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Turner MS, Simons JS, Gilbert SJ, Frith CD, Burgess PW. Distinct roles for lateral and medial rostral prefrontal cortex in source monitoring of perceived and imagined events. Neuropsychologia. 2008;46:1442–1453. doi: 10.1016/j.neuropsychologia.2007.12.029. [DOI] [PMC free article] [PubMed] [Google Scholar]
- van de Ven VG, Formisano E, Röder CH, Prvulovic D, Bittner RA, Dietz MG, … Linden DE. The spatiotemporal pattern of auditory cortical responses during verbal hallucinations. NeuroImage. 2005;27:644–655. doi: 10.1016/j.neuroimage.2005.04.041. [DOI] [PubMed] [Google Scholar]
- Vinogradov S, Luks TL, Simpson GV, Schulman BJ, Glenn S, Wong AE. Brain activation patterns during memory of cognitive agency. NeuroImage. 2006;31:896–905. doi: 10.1016/j.neuroimage.2005.12.058. [DOI] [PubMed] [Google Scholar]
- Woods RP, Cherry SR, Mazziotta JC. Rapid automated algorithm for aligning and reslicing PET images. Journal of Computer Assisted Tomography. 1992;16:620–633. doi: 10.1097/00004728-199207000-00024. [DOI] [PubMed] [Google Scholar]
- Woodward TS, Menon M. Misattribution models: II. Source monitoring in hallucinating schizophrenia subjects. In: Jardri R, Cauchia A, Thomas P, Pins D, editors. The neuroscience of hallucinations. New York, NY: Springer; 2013. pp. 169–184. [Google Scholar]