Abstract
That auditory hallucinations are voices heard in the absence of external stimuli implies the existence of endogenous neural activity within the auditory cortex responsible for their perception. Further, auditory hallucinations occur across a range of healthy and disease states that include reduced arousal, hypnosis, drug intoxication, delirium, and psychosis. This suggests that, even in health, the auditory cortex has a propensity to spontaneously “activate” during silence. Here we report the findings of a functional MRI study, designed to examine baseline activity in speech-sensitive auditory regions. During silence, we show that functionally defined speech-sensitive auditory cortex is characterized by intermittent episodes of significantly increased activity in a large proportion (in some cases >30%) of its volume. Bilateral increases in activity are associated with foci of spontaneous activation in the left primary and association auditory cortices and anterior cingulate cortex. We suggest that, within auditory regions, endogenous activity is modulated by anterior cingulate cortex, resulting in spontaneous activation during silence. Hence, an aspect of the brain's “default mode” resembles a (preprepared) substrate for the development of auditory hallucinations. These observations may help explain why such hallucinations are ubiquitous.
Keywords: auditory system, baseline activity, functional MRI
Tonic baseline activity is thought to be essentially uniform across the human brain volume (1). Specific foci (e.g., medial prefrontal regions) reduce their activity to below baseline upon commencement of structured mental processes, suggesting that an organized default mode of brain function is temporarily suspended during such processes (2). However, less is known regarding how focal baseline activity might naturally vary over time and whether episodic increases in activity might be sufficiently large to resemble statistically defined, externally evoked activation reported in functional MRI (fMRI) experiments. This question is important from both a neuroscientific and a clinical perspective, because spontaneous activation of the auditory cortex during silence is implicated in the emergence of auditory verbal hallucinations (3–6); perceptions in the absence of external stimuli, which occur in numerous states of health and disease, including reduced arousal, hypnosis, drug intoxication, delirium, and psychosis (7). This paper describes a fMRI experiment designed: (i) to examine variation of baseline activity in speech-sensitive auditory regions, in the absence of external stimuli, and (ii) to define the brain-wide functional anatomy associated with the emergence of spontaneous auditory activation during silence.
Previous work has shown that neural activity in the auditory cortex is subject to modulation by factors other than auditory sensory input. Attention to the auditory modality (8) and utilization of auditory imagery (9, 10) increase imaging parameters thought to represent markers of neural activity, including the blood oxygenation level-dependent response measured by fMRI (11). Hence, we hypothesize that auditory baseline activity is sensitive to modulation by background physiological processes that support cognitive biasing effects and, therefore, may vary in such a way that intermittent episodes of increased activity during silence are observed. The anterior cingulate cortex (ACC), part of the limbic system adjacent to the medial prefrontal cortex, is involved in the recruitment of attention (12, 13) and is a putative source of a signal that might increase baseline activity in sensory cortex (14). With respect to the auditory system, this idea is supported by evidence for coactivation of ACC and auditory regions during silence in an auditory imagery task (15) and by the widely replicated observation of ACC activation during the experience of auditory hallucinations (6, 16–20).
In the current study, our prediction (based upon the hypothesis above) is that spontaneous activation of speech-sensitive auditory cortex occurs during silence, in conjunction with activation of ACC.
Methods
This study was approved by the North and South Sheffield research ethics committees.
Voice Stimuli. To identify speech-sensitive auditory regions, we created 24 unique stimuli for use in the fMRI experiment. Each stimulus consisted of a single emotionally neutral spoken phrase of three to four words and 1- to 2-s duration [e.g., “close the door” (21–23)]. All stimuli were digitally recorded at 44.1 kHz and 16 bits and presented within the scanner (below) at ≈50 dB sound pressure level over Commander XG electrostatic headphones (Resonance Technologies, Dallas).
Subjects and Scanning Protocol. Twelve healthy right-handed [mean ± SD right-hand dominance = 75.00 ± 21.98% (24)] male subjects aged 23 ± 3 years were studied. fMRI was performed by using echo-planar imaging on a 1.5-T Eclipse system (Philips Medical Systems, Cleveland) at the University of Sheffield. To circumvent problems associated with acoustic scanner noise, we used a sparse protocol (25) utilizing haemodynamic response latency to permit stimulus presentation during silent periods (32 × 4-mm contiguous slices at 48 time points, repetition time = 20,000 ms, echo time = 50 ms, field of view = 240 mm, in-plane matrix = 128 × 128). In an alternating stimulus vs. no-stimulus design, functional brain volumes were acquired after (i) the presentation of a single voice stimulus (as described above, every 40 s) and (ii) silence, over a single 16-min run (with the gradient coils turned off, quiescent scanner room sound was effectively cancelled by 30-dB headphone attenuation). We deliberately did not utilize pseudorandom presentation of stimuli because of a concern that this might lead to systematic attentional modulation (8, 14) of the auditory baseline due to vigilance with respect to variable expectation of stimulus delivery (or nondelivery). Similarly, and unlike stimulus activation experiments, we did not use a control for attention during functional scans, because such controls are, by their very nature, demanding of attention and therefore a form of goal-directed behavior that should be excluded from baseline state investigations (1).
First-Level Image Analysis. Functional definition of speech-sensitive regions. After timing and movement correction, functional images were spatially normalized and smoothed with a Gaussian kernel of 6-mm full width at half maximum by using SPM99 (Statistical Parametric Mapping, Wellcome Department of Imaging Neuroscience, University College London, London). Scaling according to mean voxel value was applied throughout each functional volume to correct for any confounds arising from global vascular effects. The difference in blood oxygen level-dependent response between speech and silence conditions was estimated at every voxel across the whole brain volume for each individual subject by using the General Linear Model. This generated contrast images, which were interrogated to produce single-subject parametric brain maps of t statistics (P < 0.001, uncorrected) showing activation in response to our stimuli, thereby functionally defining the speech-sensitive regions (voxels) of the auditory cortex.
Estimation of baseline activity in speech-sensitive regions. The following procedure was applied separately to each cerebral hemisphere for each individual subject:
At every silence time point, the number of functionally defined voxels that were more than 2 SD above their silence-specific mean (i.e., Z > 2) was calculated.
Time points were ranked in order of total Z > 2 voxels. The time point with greatest total voxels of Z > 2 was defined as a spontaneous activation, provided this total figure exceeded 2.5% of the functionally defined speech-sensitive volume, i.e., exceeded Gaussian assumptions. Other time points were defined as spontaneous activations if they contained voxels with Z > 2 that were common to all higher-ranked time points and exceeded Gaussian assumptions with respect to the total of Z > 2 voxels. This approach utilizes both statistical height of activation (Z score) and extent of activation, i.e., a requirement that >2.5% of voxels exceed the height criterion.
Customized models were produced, defining time points of spontaneous activation during silence and time points of no such activation (i.e., the remainder of the silence baseline). Fig. 1 schematically shows the difference between a time point of spontaneous activation and the remainder of the silence baseline.
The difference in blood oxygen level-dependent signal between spontaneous activation time points (during silence) and the remainder of the silence baseline was estimated at every voxel across the whole brain volume. This procedure generated two contrast images per subject (one each for the left and right hemispheres), which were entered in second-level (group) analyses detailed below.
Second-Level Image Analyses. At the group level, we examined two main effects across the whole brain volume: left hemisphere spontaneous activation vs. remainder of silence baseline (l-spont vs. baseline) and right hemisphere spontaneous activation vs. remainder of silence baseline (r-spont vs. baseline). For each effect of interest, the relevant individual subject contrast images from the first-level analyses were entered as data points in a one-sample t test. This approach amounts to a mixed-effects analysis, with between-subject variance treated as a random effect in the statistical model. With respect to these analyses, it is important to note that our method is not a simple tautology (i.e., seeking to find foci of statistically defined activation, where activity is already known to be high), because it defines the degree of anatomical constraint that can be applied to identifying foci of spontaneous activation in the population from which the subjects were drawn (26). Also, this method allows for the main effect of spontaneous activation to be examined across the whole brain volume (i.e., both within and beyond the functionally defined speech-sensitive volume).
These second-level analyses produced group parametric maps of t statistics (P < 0.001, uncorrected). We also performed a conjoint analysis of l- and r-spont vs. baseline, to identify areas of significant overlap between those contrasts.
Spatial normalization in SPM99 (above) produced interpolated 2-mm3 voxels; for the purposes of reporting and neuroanatomical labeling, all stereotactic coordinates were transformed into the space described by Talairach and Tournoux (27).
Results
Spontaneous Activation of Speech-Sensitive Regions During Silence. Temporal domain. Each subject demonstrated evidence of spontaneous auditory activation during silence in both hemispheres. Table 1 shows the mean speech-sensitive and observed spontaneous activation voxel data and also an estimate of the expected spontaneous activation data (for the example, mean, case) from the binomial distribution.
Table 1. Mean functionally defined speech-sensitive voxels in 12 subjects and expected and observed voxels with Z > 2 during specified time points of spontaneous activation in the left and right hemispheres.
Left | Right | |
---|---|---|
Mean functionally defined voxels | 790 | 779 |
Mean ± SD expected Z > 2 voxels | 19.75 ± 4.39 | 19.48 ± 4.36 |
Mean observed Z > 2 voxels | 96 | 81 |
The probability of observing these Z > 2 data under Gaussian assumptions, from the normal approximation to the binomial distribution, is very small (<1 × 10-6).
In the left hemisphere, 11 subjects exhibited two time points of spontaneous activation (8.33% of sampled baseline time), and one subject exhibited three such time points (12.50% of sampled baseline time). The mean ± SD percentage of speech-sensitive voxels with Z > 2 during spontaneous activation was 12.41 ± 8.26%. Example data from a single subject are shown in Fig. 2.
In the right hemisphere, eight subjects demonstrated two time points of spontaneous activation (8.33% of sampled baseline time), and four subjects demonstrated a single time point of spontaneous activation (4.17% of sampled baseline time). The mean ± SD percentage of speech-sensitive voxels with Z > 2 during spontaneous activation was 9.52 ± 6.16%.
Of the 35 time points of spontaneous activation occurring in at least one hemisphere, 42% involved only the left hemisphere, 29% involved only the right hemisphere, and 29% simultaneously involved both the left and right hemispheres.
To check for any temporal grouping of those time points where spontaneous activation was observed, we divided the functional scanning sessions into quartile time bins, each containing six silence time points. The occurrence of spontaneous activation time points did not significantly differ among the temporal quartiles (χ2 = 3.27; P = 0.35).
Because images were smoothed with a Gaussian kernel during preprocessing, which might accentuate spatial correlations in the data, we also performed these analyses using unsmoothed images. This did not result in appreciably different percentages of speech-sensitive voxels with Z > 2 during spontaneous activation and did not affect the significance of the difference between that expected and observed (Table 1). By way of a further sensitivity analysis, we reanalyzed our data assuming a Rayleigh distribution, which has a heavy positive tail and is thought to describe the worst-case scenario where reconstructed MRI amplitude images consist only of noise (28). This did not materially affect our results.
Spatial domain. Contrasting time points of spontaneous activation in the left hemisphere with the remainder of the silence baseline [l-spont vs. baseline; Fig. 3 (yellow)] revealed foci in the left lateral superior temporal gyrus [(STG); Brodmann's area (BA) 22; coordinates: –55, –11, 4; 145 voxels exceeded P < 0.001, uncorrected; t = 9.22) and more posterior STG (BA 22; coordinates: –44, –31, 11; 129 voxels; t = 7.51), overlapping with the transverse temporal gyrus (TTG; BA 41). Furthermore, the r-spont vs. baseline contrast [Fig. 3 (red)] also revealed foci in the left STG (BA 22; coordinates: –57, –38, 18; 63 voxels; t = 7.72) and TTG (BA 41; coordinates: –36, –28, 14; 52 voxels; t = 6.63); significant overlap between l- and r-spont conditions in these regions was confirmed by conjoint statistical analysis [t = 9.19 (STG); t = 7.69 (TTG)].
In the l-spont vs. baseline contrast, we found a single further focus in the left ACC (BA 32; coordinates: –2, 53, 12; 56 voxels; t = 6.56), which extended to include part of the medial frontal gyrus (BA 9/10). The glass brain in Fig. 4 demonstrates the specificity of this result.
Further Analyses. We repeated the main analysis, substituting Z < –2 for Z > 2 to identify time points of spontaneous deactivation during silence. Using a within-subject repeated measures ANOVA, we compared total percentage Z > 2 voxels during spontaneous activation (reflecting both number of time points and voxels) with Z <–2 voxels during spontaneous deactivation. In the left hemisphere, more voxels achieved Z > 2 during spontaneous activation than Z < –2 during spontaneous deactivation, but this effect was not observed in the right hemisphere (Table 2). Hence, there was a significant direction (positive vs. negative) by hemisphere interaction (F = 5.79; P = 0.03).
Table 2. Median time points and mean total percentage speech-sensitive voxels for spontaneous activation and deactivation and unexpectedly large stimulus-evoked activation.
Condition | Criterion | Median (range) time points | Mean ± SD total % voxels |
---|---|---|---|
Spontaneous activation | |||
Left hemisphere | Z > 2 | 2 (2-3) | 25.85 ± 13.27 |
Right hemisphere | 2 (1-2) | 15.87 ± 8.61 | |
Spontaneous deactivation | |||
Left hemisphere | Z < -2 | 1.5 (1-2) | 15.13 ± 5.40 |
Right hemisphere | 1 (1-2) | 16.31 ± 9.05 | |
Unexpectedly large stimulus-evoked activation | |||
Left hemisphere | Z > 2 | 1 (1-3) | 17.66 ± 14.24 |
Right hemisphere | 2 (1-2) | 17.85 ± 11.11 |
The mean total percentage of voxels with Z > 2 (or < -2) reflects both numbers of time points and voxels with Z > 2 (< -2) at such time points, for each subject (n = 12). Note that the mean total percentage of voxels with Z > 2 (< -2) is relatively homogeneous across conditions and hemispheres, with the exception of spontaneous activation in the left hemisphere, where more voxels are involved (see text for statistical analysis).
We also examined variation in magnitude of stimulus-evoked activation by repeating the main analysis using data from the speech (not silence) time points and a criterion of Z > 2 with respect to the speech-specific mean. This procedure identified time points with more Z > 2 voxels than would be expected under Gaussian assumptions, which we refer to as time points of unexpectedly large activation. More left-hemisphere voxels achieved Z > 2 during spontaneous activation (in silence) than Z > 2 during unexpectedly large stimulus-evoked activation; this was not observed in the right hemisphere (Table 2). A condition (silence vs. speech) by hemisphere interaction was significant (F = 5.39; P = 0.04).
In the entire group of subjects, across all brain voxels and all silence time points, Z score distribution closely approximated to a standard Gaussian distribution (Kolmogorov–Smirnov statistic = 0.005; P > 0.05). Repeating our main analysis with data from the whole brain (i.e., without the functional mask of speech-sensitive voxels) revealed some time points of spontaneous activation occurring anywhere in the brain during silence. However, the mean ± SD percentage of voxels with Z > 2 was small (3.86 ± 1.66%). In particular, this percentage was significantly smaller than in the case of left- (t = 5.18; P < 0.001) and right- (t = 4.50; P < 0.001) hemisphere spontaneous activation of speech-sensitive regions. Furthermore, time points of spontaneous activation in auditory regions and anywhere in the brain were not usually identical (80% of left- and 75% of right-hemisphere time points of spontaneous activation in speech-sensitive regions were distinct). Application of the (spatial) mixed-effects analysis revealed that time points of spontaneous activation occurring anywhere in the brain were associated with two foci in the right insula. Hence, there was evidence for perturbation of the overarching Gaussian distribution of Z scores, which we observed across the whole brain volume, at those foci (coordinates: 44, 17, –1; 31 voxels exceeded P < 0.001, uncorrected; t = 7.99 and coordinates: 42, –2, 2; 30 voxels; t = 5.21).
Finally, to exclude the possibility that our results were influenced by any random walks of the MRI scanner field, we also acquired and analyzed data from a phantom head with internal structure. We did not detect evidence of spontaneous activation.
Discussion
We have shown that temporal variation in the auditory baseline is manifest as intermittent episodes of strikingly increased activity within speech-sensitive regions. These episodes arise in the absence of external auditory stimulation and exceed the statistical definition of activation used in functional image analysis. Hence, we have referred to such phenomena as spontaneous activations. Anatomical foci of spontaneous activation are located in the left STG and medial TTG [site of the primary auditory cortex in humans (29)].
These data suggest that spontaneous auditory activation in both left and right hemispheres can be explained by foci in the left temporal cortex (Fig. 3). Essentially, this means that bilateral spontaneous activation (in the time domain) implicates a specific population of left-hemisphere voxels (in the space domain) that is consistent between subjects. On the other hand, right-hemisphere spontaneous activation (in the time domain) is associated with greater intersubject variability with respect to the right-hemisphere voxels that are involved, but simultaneous intersubject consistency with respect to a population of left-hemisphere voxels (associated with right-hemispheric activation). Overall, we argue these results suggest that spontaneous activation of speech-sensitive regions is driven by a left-hemispheric effect, which impacts upon the right hemisphere via homotopic connections (30, 31). This notion can be further specified by considering similarities and differences in topography between left-hemisphere foci associated with left- and right-hemisphere spontaneous activation. Voxels in the left-lateral STG are associated only with left-hemisphere effects, but voxels in the left medial TTG and posterior STG are associated with left and right spontaneous activation phenomena (Fig. 3). Hence, these latter foci are more likely to be the sources of effects that can be detected in the right hemisphere, perhaps with the posterior STG as predominant due to its involvement in processing complex characteristics of speech (21).
Our analysis of spontaneous deactivation also suggests that spontaneous auditory activation is a left-lateralized phenomenon. Although the data show that Z score distribution in the auditory baseline is somewhat more polarized than throughout the entire brain (Table 2), with more voxels achieving Z > 2 and Z <–2 than would be expected under Gaussian assumptions, we can nonetheless detect evidence of spontaneous activation above and beyond any such polarization. Specifically, more voxels achieve Z > 2 than Z < –2 during silence, but only in the left hemisphere. Similarly, more voxels achieve Z > 2 during silence than Z > 2 during speech perception (i.e., by comparison with their speech-specific mean) but only in the left hemisphere (Table 2). The latter observation is compatible with speech serving to constrain spontaneous activation in left-sided speech-sensitive regions; an interpretation consistent with the idea that externally evoked activation and endogenous activation may compete for finite neural resources up to a point of saturation (4).
The observed lateralization of spontaneous activation in speech-sensitive auditory cortex is in accordance with many demonstrations of left-hemispheric dominance for speech and language functions (32). It is possible that such dominance derives from interhemispheric differences in neuronal properties such as connectivity patterns and strengths and synaptic gain control, caused by asymmetries in the regional distribution of neurotransmitters (or their receptors), or electrophysiological differences related to the distribution of specific ion channels and transporters (33–37). These mechanisms might also account for the left auditory cortex's propensity to undergo spontaneous (and anatomically focal) activation during silence. Furthermore, although global vascular effects on the blood oxygen level-dependent signal have been shown to demonstrate some regional variation (38), the distinct lateralization of spontaneous activation that we observed (together with our correction for mean voxel value) provides evidence against any such vascular confound in the current data.
As predicted, the ACC was revealed as an extratemporal focus where activity was associated with spontaneous activation of left hemisphere speech-sensitive auditory cortex. This finding is consistent with ACC exerting influence over regionally specific brain activity (12), a process that might amount to the physiological basis of attention (13, 14). Given the overlap between our results and those of an auditory imagery experiment (15), it could be tempting to suggest that we have imaged the correlates of attention to thoughts involving the auditory domain. However, we deliberately sought to examine baseline auditory processes without reference to any specific cognitive task or mental state. Ours is a study of auditory physiology, not phenomenology. Hence, we avoid any dualist interpretation of our results, instead emphasizing that the physiological processes that we have described might represent a generic mechanism by which the brain supports modulation of focal auditory activity [evident in cognition experiments (8–10)].
Moreover, if an aspect of the brain's default mode (1, 2) is well suited to support physiological modulation of auditory activity, then it is also possible that the brain might be naturally predisposed to sustain aberrant auditory activity in pathological states, manifest as auditory hallucinations. We believe this is an important issue, because attempts to define a structural abnormality in the auditory cortices of people with auditory hallucinations have been subject to problems of nonreplication (39, 40). Our results suggest that the concept of a focal anatomical lesion need not necessarily be invoked to explain auditory hallucinations. Rather, these states might arise from a functional disturbance of the auditory system, which shows an unstable baseline even in healthy subjects. Fluctuation of activity within the temporocingulate network that we have described is associated with nonperception in health, but it is possible that dysfunction within such a network could lead to false perception in disease (3–6). Such transformation into perception could arise under conditions of heightened ACC influence on auditory cortical function, a notion consistent with the widely replicated finding of ACC activation during auditory hallucinations (6, 16–20).
Outside the mask derived from functional definition of speech-sensitive voxels, the Z score distribution across the dimensions of whole brain space and scan time was close to Gaussian. Nonetheless, we were able to identify foci in the right insula that, like the left temporal cortex, were associated with spontaneous activation (but at different time points). Although these findings are of their own interest [the functional associations of the insula are numerous (41)], they also enable further specification of our main results. First, it is apparent that time points of defined spontaneous activation detectable within the speech-sensitive and whole-brain volumes are not (generally) the same. Second, even if they were the same, the percentage of voxels with Z > 2 anywhere in the brain would be insufficiently large to explain the speech-sensitive voxel data. Hence, the functional mask of speech-sensitive voxels is not a window on more widespread phenomena, and spontaneous auditory activation is not simply an expression of fluctuating Z scores across the whole brain. Moreover, it is possible (indeed, a further testable hypothesis) that the overarching Gaussian distribution of Z scores reflects baseline activity in numerous functionally defined subvolumes of the brain (e.g., auditory, visual, and somatosensory regions), each with the capacity to undergo spontaneous activation, but at different time points.
Although we have used the term spontaneous activation, this phenomenon differs from the concept of activation (within a deterministic system) that is commonly used in models of neuroimaging data (42, 43). In such models, the brain can be activated only by driving external inputs, i.e., experimental stimuli. On the other hand, in this study, we have described a situation where structured activation is generated by the brain itself. This demonstrates one of the major discrepancies between neurocomputational models and physiological observations: the presence of nonrandom internally generated information derived from physiological substrata such as oscillating ion channels, which are found at high concentrations within distinct generator regions (44–46). We thus suggest that the neuroimaging concept of activation could be expanded to include the effects of spontaneously discharging internal generators. In this context, it is important to emphasize that spontaneous is not equivalent to random; the shifts in baseline activity are distinctly clustered in both the temporal (Fig. 2) and spatial (Figs. 3 and 4) domains (i.e., relatively proximal voxels exhibit unusually high signal levels at similar times).
Conclusion
We have provided evidence for a phenomenon of spontaneous activation in the baseline of speech-sensitive auditory regions. Such activation arises in the left primary auditory cortex (TTG) and association auditory cortex (STG), in conjunction with activation of the ACC. If, as we hypothesize, this temporocingulate network represents a generic mechanism by which the brain supports modulation of focal auditory activity, then dysfunction within such a network could be manifest as auditory hallucinations. Hence, we propose a biologically informed model for auditory hallucinations based upon perturbation of normal baseline processes, a model that might explain why auditory hallucinations are ubiquitous.
Acknowledgments
The Wellcome Trust supported this work.
Author contributions: M.D.H., S.B.E., and P.W.R.W. designed research; M.D.H., T.W.R.M., and I.D.W. performed research; M.D.H., S.B.E., T.W.R.M., T.F.D.F., and I.D.W. analyzed data; and M.D.H., S.B.E., T.F.D.F., I.D.W., and P.W.R.W. wrote the paper.
Conflict of interest statement: No conflicts declared.
This paper was submitted directly (Track II) to the PNAS office.
Abbreviations: ACC, anterior cingulate cortex; BA, Brodmann's area; fMRI, functional MRI; STG, superior temporal gyrus; TTG, transverse temporal gyrus.
References
- 1.Raichle, M. E., MacLeod, A. M., Snyder, A. Z., Powers, W. J., Gusnard, D. A. & Shulman, G. L. (2001) Proc. Natl. Acad. Sci. USA 98, 676–682. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Gusnard, D. A. & Raichle, M. E. (2001) Nat. Rev. Neurosci. 2, 685–694. [DOI] [PubMed] [Google Scholar]
- 3.Woodruff, P., Brammer, M., Mellers, J., Wright, I., Bullmore, E. & Williams, S. (1995) Lancet 346, 1035. [DOI] [PubMed] [Google Scholar]
- 4.Woodruff, P. W. R, Wright, I. C., Bullmore, E. T., Brammer, M., Howard, R. J., Williams, S. C. R., Shapleske, J., Rossell, S., David, A. S., McGuire, P. K., et al. (1997) Am. J. Psychiatry 154, 1676–1682. [DOI] [PubMed] [Google Scholar]
- 5.Dierks, T, Linden, D. E. J., Jandl, M., Formisano, E., Goebel, R., Lanfermann, H. & Singer, W. (1999) Neuron 22, 615–621. [DOI] [PubMed] [Google Scholar]
- 6.Shergill, S. S., Brammer, M. J., Williams, S. C. R., Murray, R. M. & McGuire, P. K. (2000) Arch. Gen. Psychiatry 57, 1033–1038. [DOI] [PubMed] [Google Scholar]
- 7.Woodruff, P. W. R. (2004) Cognit. Neuropsychiatry 9, 73–91. [DOI] [PubMed] [Google Scholar]
- 8.Woodruff, P. W., Benson, R. R., Bandettini, P. A., Kwong, K. K., Howard, R. J., Talavage, T., Belliveau, J. & Rosen, B.R. (1996) NeuroReport 7, 1909–1913. [DOI] [PubMed] [Google Scholar]
- 9.Halpern, A. R., Zatorre, R. J., Bouffard, M. & Johnson, J. A. (2004) Neuropsychologia 42, 1281–1292. [DOI] [PubMed] [Google Scholar]
- 10.Kraemer, D. J. M., Macrae, C. M., Green, A. E. & Kelley, W. M. (2005) Nature 434, 158. [DOI] [PubMed] [Google Scholar]
- 11.Logothetis, N. K., Pauls, J., Augath, M., Trinath, T. & Oeltermann, A. (2001) Nature 412, 150–157. [DOI] [PubMed] [Google Scholar]
- 12.Bush, G., Luu, P. & Posner, M. I. (2000) Trends Cognit. Sci. 4, 215–222. [DOI] [PubMed] [Google Scholar]
- 13.Stephan, K. E., Marshall, J. C., Friston, K. J., Rowe, J. B., Ritzl, A., Zilles, K. & Fink, G. R. (2003) Science 301, 384–386. [DOI] [PubMed] [Google Scholar]
- 14.Chawla, D., Rees, G. & Friston, K. J. (1999) Nat. Neurosci. 2, 671–676. [DOI] [PubMed] [Google Scholar]
- 15.McGuire, P. K., Silbersweig, D. A., Murray, R. M., David, A. S., Frackowiak, R. S. J. & Frith, C. D. (1996) Psychol. Med. 26, 29–38. [DOI] [PubMed] [Google Scholar]
- 16.Cleghorn, J. M., Garnett, E. S., Nahmias, C., Brown, G. M., Kaplan, R. D., Szechtman, H., Szechtman, B., Franco, S., Dermer, S. W. & Cook, P. (1990) Br. J. Psychiatry 157, 562–570. [DOI] [PubMed] [Google Scholar]
- 17.Cleghorn, J. M., Franco, S., Szechtman, B., Kaplan, R. D., Szechtman, H., Brown, G. M., Nahmias, C. & Garnett, E. S. (1992) Am. J. Psychiatry 149, 1062–1069. [DOI] [PubMed] [Google Scholar]
- 18.McGuire, P. K., Shah, G. M. S. & Murray, R. M. (1993) Lancet 342, 703–706. [DOI] [PubMed] [Google Scholar]
- 19.Silbersweig, D. A., Stern, E., Frith, C., Cahill, C., Holmes, A., Grootoonk, S., Seaward, J., McKenna, P., Chua, S. E., Schnorr, L., et al. (1995) Nature 378, 176–179. [DOI] [PubMed] [Google Scholar]
- 20.Szechtman, H., Woody, E., Bowers, K. S. & Nahmias, C. (1998) Proc. Natl. Acad. Sci. USA 95, 1956–1960. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Hunter, M. D., Griffiths, T. D., Farrow, T. F. D., Zheng, Y., Wilkinson, I. D., Hegde, N., Woods, W., Spence, S. A. & Woodruff, P. W. R. (2003) Brain 126, 161–169. [DOI] [PubMed] [Google Scholar]
- 22.Hunter, M. D., Smith, J. K., Taylor, N., Woods, W., Spence, S. A., Griffiths, T. D. & Woodruff, P. W. R. (2003) Percept. Mot. Skills 97, 246–250. [DOI] [PubMed] [Google Scholar]
- 23.Hunter, M. D., Phang, S. Y., Lee, K.-H. & Woodruff, P. W. R. (2005) Neurosci. Lett. 375, 148–150. [DOI] [PubMed] [Google Scholar]
- 24.Oldfield, R. C. (1971) Neuropsychologia 9, 97–113. [DOI] [PubMed] [Google Scholar]
- 25.Hall, D. A., Haggard, M. P., Akeroyd, M. A., Palmer, A. R., Summerfield, A. Q., Elliot, M. R., Gurney, E. M. & Bowtell, R. W. (1999) Hum. Brain Mapp. 7, 213–223. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Friston, K. J., Holmes, A. P. & Worsley, K. J. (1999) NeuroImage 10, 1–5. [DOI] [PubMed] [Google Scholar]
- 27.Talairach, P. & Tournoux, J. A. (1988) A Stereotactic Co-planar Atlas of the Human Brain (Thieme, Stuttgart).
- 28.Sijbers, J., den Dekker, A. J., Van Audekerke, J., Verhoye, M. & Van Dyck, D. (1998) Magn. Reson. Imaging 16, 87–90. [DOI] [PubMed] [Google Scholar]
- 29.Rademacher, J., Morosan, P., Schleicher, A., Freund, H.-J. & Zilles, K. (2001) NeuroReport 12, 1561–1565. [DOI] [PubMed] [Google Scholar]
- 30.Woodruff, P. W. R., Pearlson, G. D., Geer, M. J., Barta, P. E. & Chilcoat, H. D. (1993) Psychol. Med. 23, 45–56. [DOI] [PubMed] [Google Scholar]
- 31.Woodruff, P. W. R., Phillips, M. L., Rushe, T., Wright, I. C., Murray, R. M. & David, A. S. (1997) Schizophr. Res. 23, 189–196. [DOI] [PubMed] [Google Scholar]
- 32.Scott, S. K., Blank, C. C., Rosen, S. & Wise, R. S. J. (2000) Brain 123, 2400–2406. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Belin, P., Zilbovicius, M., Crozier, S., Thivard, L., Fontaine, A., Masure, M. C. & Samson, Y. (1998) J. Cognit. Neurosci. 10, 536–540. [DOI] [PubMed] [Google Scholar]
- 34.Morand, N., Bouvard, S., Ryvlin, P., Mauguiere, F., Fischer, C., Collet, L. & Veuillet, E. (2001) Acta Otolaryngol. 121, 293–296. [DOI] [PubMed] [Google Scholar]
- 35.Friederici, A. D. & Alter, K. (2004) Brain Lang. 89, 267–276. [DOI] [PubMed] [Google Scholar]
- 36.Pandya, P. K., Moucha, R., Engineer, N. D., Rathbun, D. L., Vazquez, J. & Kilgard, M. P. (2005) Hear. Res. 203, 10–20. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Philibert, B., Beitel, R. E., Nagarajan, S. S., Bonham, B. H., Schreiner, C. E. & Cheung, S. W. (2005) J. Comp. Neurol. 487, 391–406. [DOI] [PubMed] [Google Scholar]
- 38.Dagli, M. S., Ingeholm, J. E. & Haxby, J. V. (1999) NeuroImage 9, 407–415. [DOI] [PubMed] [Google Scholar]
- 39.Shapleske, J., Rossell, S. L., Woodruff, P. W. R. & David, A. S. (1999) Brain Res. Rev. 29, 26–49. [DOI] [PubMed] [Google Scholar]
- 40.Shapleske, J., Rossell, S. L., Simmons, A., David, A. S. & Woodruff, P. W. R. (2001) Biol. Psychiatry 49, 685–693. [DOI] [PubMed] [Google Scholar]
- 41.Augustine, J. R. (1996) Brain Res. Brain Res. Rev. 22, 229–244. [DOI] [PubMed] [Google Scholar]
- 42.Friston, K. J., Harrison, L. & Penny, W. (2003) NeuroImage 19, 1273–1302. [DOI] [PubMed] [Google Scholar]
- 43.David, O., Harrison, L. & Friston, K. J. (2005) NeuroImage 25, 756–770. [DOI] [PubMed] [Google Scholar]
- 44.Buzsaki, G. & Draguhn, A. (2004) Science 304, 1926–1929. [DOI] [PubMed] [Google Scholar]
- 45.Kuhlman, S. J. & McMahon, D. G. (2004) Eur. J. Neurosci. 20, 1113–1117. [DOI] [PubMed] [Google Scholar]
- 46.Ramirez, J. M., Tryba, A. K. & Pena, F. (2004) Curr. Opin. Neurobiol. 14, 665–674. [DOI] [PubMed] [Google Scholar]