Abstract
Inferences about the taste of foods are a key aspect of our everyday experience of food choice. Despite this, gustatory mental imagery is a relatively under-studied aspect of our mental lives. In the present study, we examined subjects during high-field fMRI as they actively imagined basic tastes and subsequently viewed pictures of foods dominant in those specific taste qualities. Imagined tastes elicited activity in the bilateral dorsal mid-insula, one of the primary cortical regions responsive to the experience of taste. In addition, within this region we reliably decoded imagined tastes according to their dominant quality - sweet, sour, or salty – thus indicating that, like actual taste, imagined taste activates distinct quality-specific neural patterns. Using a cross-task decoding analysis, we found that the neural patterns for imagined tastes and food pictures in the mid-insula were reliably similar and quality-specific, suggesting a common code for representing taste quality regardless of whether explicitly imagined or automatically inferred when viewing food. These findings have important implications for our understanding of the mechanisms of mental imagery and the multimodal nature of presumably primary sensory brain regions like the dorsal mid-insula.
Keywords: Insula, Taste, Cognition, Mental Imagery, MVPA
INTRODUCTION
Food-related decision making is a frequent and integral part of our daily life. Many different factors influence our decisions on what foods to eat, including cost, healthiness, and importantly, taste. When deciding between alternate foods, we often imagine what these foods will taste like. This form of explicit gustatory imagery draws upon our past experiences with similar foods to inform our decision making. Neuroscientific studies of food choice, based on decades of research in rodent and non-human primate models, generally focus on the decision-making process itself, and how it is guided by the function of the brain’s hedonic and reward circuitry (see Berridge and Kringelbach, 2015; Rudebeck and Murray, 2014 for reviews). Many of the more subjective aspects of food choice that play a central role in everyday food choice in humans, including the role of gustatory imagery, are surprisingly understudied.
Recent neuroimaging studies have revealed taste-related activation with the dorsal mid-insula, located in the fundus of the central insular sulcus and the overlying frontoparietal operculum, assumed to be the primary cortical region for taste (Small, 2010; Veldhuizen et al., 2011, Yeung et al., 2017). High-resolution gustatory fMRI studies, using a combination of univariate and multivariate analyses, have shown that distinct taste qualities are represented in this region by distributed spatial patterns rather than distinct topographic areas (Avery et al., 2020; 2021b; Chikazoe et al., 2019). These results suggest that taste quality is represented in the insular cortex by a spatial population code, a result supported by recent calcium imaging studies in the rodent insula (Fletcher et al., 2017; Levitan et al., 2019; Chen et al., 2020).
It has also been recently shown that the response to food pictures within the dorsal mid-insula could be classified based on the dominant taste category of the depicted foods (Avery et al., 2021a). This result showed that merely viewing pictures of foods triggers an automatic retrieval of taste property information within this region, a representation which is fine-grained enough to distinguish between images of foods based on their primary associated taste (i.e., sweet, salty, sour). They were also consistent with the view that that higher-order inferences derived from stimuli in one modality (i.e., vision) could be represented in brain regions typically thought to represent only low-level information about a different modality (i.e., taste).
However, although both actual tastes (sweet, salty, sour) delivered during scanning and food images classified by their dominant taste (candy, pretzels, lemons) could be decoded from the same dorsal mid-insula region in this previous study (Avery et al., 2021a), cross-decoding tastes and images was unsuccessful, as their associated patterns of response were not reliably similar. One possibility for this result is that viewing pictures elicits inferences about other food-related properties than just taste, such as shape, color, and texture. Indeed, the taste category of food pictures was also decoded in widespread regions of ventral occipitotemporal cortex, areas typically associated with these other object properties (Kravitz et al., 2013). This suggests that the difference in the neural response across food picture categories in the dorsal mid-insula could be due to other factors besides the dominant taste of the depicted foods.
Another possibility that cross-decoding was unsuccessful is overt differences between the two task paradigms, such as head motion due to swallowing, or simply that experienced tastes and inferred responses to food pictures activate different neural populations within the mid-insula, even when representing the same information. One way to circumvent this latter possibility would be to have individuals actively imagine a specific taste quality and then compare the evoked response to food pictures to the patterns of those imagined tastes. As the inferred responses to food pictures in the dorsal mid-insula function essentially as automatic and obligatory gustatory mental imagery, they would potentially share a common neural coding scheme with explicit gustatory mental imagery. Accordingly, previous studies have demonstrated successful cross-decoding of object-related sounds and mental imagery content in early visual cortex (Vetter et al., 2014). In the present study, we sought to determine if requiring subjects to imagine specific tastes could also activate the dorsal mid-insula, and if so, whether the dominant taste quality associated with food pictures could be decoded within this region using the patterns of those imagined tastes. If so, this would provide compelling evidence for the dorsal mid-insula’s role in the representation of imagined and inferred tastes in a region of the brain that also responds to experienced taste quality.
RESULTS
Behavioral Results:
In total, 22 participants completed the Taste Imagining task during fMRI, 21 completed the Food Pictures task, and 20 completed both tasks within a single session. Participants reported an average hunger rating of 3.3(2.1), an average thirst rating of 4.3(1.9), an average fullness rating of 5.2(1.9), and an average tiredness rating of 3.8(1.6). Participants reported having last eaten an average of 3.3(4.1) hours prior to the scan session. Participants first performed an fMRI task in which they imagined the taste of various substances (sugar, salt, lemon juice, and water) during scanning (Figure 1a). During post-scan follow-up questions, 8 participants reported no difficulty imagining any of the tastes, 4 indicated water was most difficult to imagine, 3 indicated sugar, 3 indicated salt, and 4 indicated lemon juice. Participants next performed a task where they saw pictures of a variety of sweet, sour, and salty foods (e.g., candy, limes, pretzels) as well as nonfood objects (Fig. 1b and Materials and Methods). During this task, subjects detected picture repetitions with an average detection accuracy of 92.1%.
Figure 1:

Experimental Design. A. Participants first performed the Taste Imagining task, within which they were instructed to imagine the taste of a specific substance indicated by the word on the center of the display screen. B. They next performed the Food Pictures fMRI task, within which they viewed pictures of a variety of food and nonfood objects within randomly ordered presentation blocks during scanning. Foods within this task were categorized into predominantly sweet (ice cream, cookies, honey, donuts), salty (chips, french fries, saltines, pretzels), and sour (lemon, lime, lemon candy, grapefruit) foods, as well as nonfood familiar objects (e.g., basketballs, tennis balls, lightbulbs, baseball gloves; see Figure S1 for examples of each image stimulus).
Imaging Results: Univariate
Taste Imagine Task:
We identified bilateral clusters in the dorsal mid-insular cortex that exhibited a greater response to the imagined tastes of sugar, salt, and lemon juice than the imagined taste of water (Figure 2, Table S1). Beyond the insula, we also observed significant response to imagined tastes > imagined water in post-central gyrus, in the approximate area of the oral somatosensory cortex (Figure 2; see Table S1 for full list of clusters).
Figure 2:

A) Bilateral regions of the dorsal mid-insular cortex and the left oral somatosensory cortex are responsive to the imagined taste of sugar, salt, and lemon juice, vs. the imagined taste of water. B) Statistical maps projected onto an inflated cortical surface model. Statistical maps were thresholded at p < 0.001 voxelwise, with a cluster-size correction for multiple comparisons at p-FWE < 0.05.
Food Pictures Task:
We identified a significant response for all food vs. object pictures in bilateral regions of the mid-insula and ventral anterior insula (Figure S2, Table S1), consistent with prior univariate results using this task (Avery et al., 2021a). Viewing food pictures, relative to object pictures, also led to greater activation of widespread areas of visual cortex. In contrast, object pictures, relative to food pictures, elicited greater activation of multiple areas in ventral occipito-temporal cortex, as well as the right dorsal anterior insular cortex (Figure S2, Table S1).
Imaging Results: Multivariate
Taste Imagine Task:
Our multivariate searchlight analysis identified significant above-chance accuracy for classifying between imagined sugar, salt, and lemon juice within the left dorsal mid-insula as well as the left dorsal and ventral anterior insula (Figure 3a, Table S2). Multiple additional regions of the brain, including the left inferior frontal gyrus, left oral somatosensory cortex, medial prefrontal cortex, and multiple visual cortical areas also exhibited significant and above chance classification accuracy for discriminating between imagined tastes (see Table S2 for a comprehensive list of clusters).
Figure 3:

MVPA Searchlight results. A) MVPA searchlights reliably classify imagined tastes (sugar, salt, lemon juice) within brain regions including the left mid-insula and oral somatosensory cortex; B) Food picture category was reliably classified within bilateral dorsal mid-insula and orbitofrontal cortex; and C) A conjunction of both searchlight maps (projected onto an inflated cortical surface model) identifies a set of regions, including the left dorsal mid-insula and ventral occipito-temporal cortex, which reliably discriminate between imagined tastes as well as food picture category. Statistical maps were thresholded at p < 0.001 voxelwise, with a cluster-size correction for multiple comparisons at p-FWE < 0.05.
Food Pictures Task:
In keeping with the multivariate results of our previous study (Avery et al., 2021a), we were able to classify food pictures based on their dominant taste quality within the bilateral dorsal mid-insular cortex, bilateral orbitofrontal cortex (BA11l) (Figure 3b, Table S2). As in that previous study, we were also able to classify the food pictures by their taste quality within large areas of visual and occipito-temporal cortex (Figure 3b; See Table S2 for comprehensive list of clusters). As an important control analysis, we were also able to classify non-food object pictures from this task within ventral occipito-temporal brain regions (Figure S3), but not in any brain regions outside of those areas.
A conjunction of the multivariate searchlight maps from both the Taste Imagining and Food Pictures task identified the left dorsal mid-insula, left orbitofrontal cortex, and throughout ventral occipito-temporal cortex (Figure 3c).
Cross-classification Analyses:
Lastly, we performed a cross-classification searchlight, training on the response to the imagined tastes of sugar, salt, and lemon juice from the taste imagining task and testing on the response to pictures of sweet, salty, and sour foods from the food pictures task, and vice versa. Our whole-brain searchlight, FWE-corrected for multiple comparisons, specifically identified the bilateral dorsal mid-insula - an area where we were also able to classify experienced tastes (Avery et al., 2020, 2021) - as the only region where imagined tastes and food pictures could be reliably cross-classified (Figure 4a, Table S2). We performed confirmatory region-of-interest (ROI) analyses using the taste-responsive insula regions from our previous study of taste decoding (Avery et al., 2020). We were able to successfully cross-classify imagined tastes and food pictures in both directions within the bilateral dorsal mid-insula (Figure 4b, Table S3).
Figure 4:

A) Cross-task classification analyses identify bilateral regions of the dorsal mid-insular cortex wherein the taste category of food pictures can reliably predicted by training on imagined tastes, and vice versa (left mid-insula pictured here). The statistical map was thresholded at p < 0.005 voxelwise, with a cluster-size correction for multiple comparisons at p-FWE < 0.05. (B & C) Classification analyses within an independent, taste-responsive region-of-interest from a previous gustatory mapping study (Avery et al., 2020) confirm the reliable cross-classification of imagined tastes and food pictures. (D) Pattern similarity analyses in this ROI identify reliably similar multi-voxel patterns for imagined tastes and food pictures. (E) Tests of ON vs. OFF diagonal similarity across subjects demonstrate that the similarity of those patterns is selective to the specific taste quality. ** p < 0.01, ** p < 0.001.
Using a pattern-similarity analysis within these ROIs, we also identified that the multivoxel patterns for imagined tastes and food pictures were reliably similar (Figure 4 d,e), and highly specific, as the on-diagonal similarity within taste-quality was significantly greater than off-diagonal similarity between taste qualities (p < 0.01, Figure 4e). This indicates that, on average, the responses to imagined tastes were more similar to the responses to their matched food picture category (e.g., imagined sugar and sweet foods) than to the non-matched food picture category.
Additional ROI Analyses:
We performed an additional series of ROI analyses to examine the effects on classification accuracy of hemispheric laterality, classification type (within-task or between-task), self-reported hunger, and difficulties in imagining any of the tastes. Firstly, our task (food decoding, imagined taste decoding, and cross-decoding) by region (Left mid-insula, Right mid-insula) ANOVA identified a significant effect of task (F = 6.1; p < 0.001), a significant effect of region (F = 13.7; p < 0.001), but no region-by-task interaction (F = 1.3; p = 0.27), which signifies that decoding accuracy was overall greater in the left than the right dorsal mid-insula. We identified a significant effect of classification type (F = 12.4, p < 0.001), indicating that within-task classification accuracy was significantly greater than between-task accuracy. We observed no significant effect of hunger overall (p = 0.72) and no significant interactions of hunger-by-task (p=0.11), hunger-by-region (p=0.83), or hunger-by-task-by-region (p=0.59).
We did not observe any overall effect on hemodynamic response due to difficulty imagining a taste during the Taste Imagining task (p = 0.27), nor did we observe any interactions of difficulty-by-condition (p=0.37), difficulty-by-region (p=0.34), or difficulty-by-condition-by-region (p=0.40). Likewise, we did not observe any difference in classification accuracy for the Taste Imagining task in the left mid-insula between subjects that reported difficulty imagining any of the tastes (n = 9) and those that did not (n = 11; t = −0.39, p=0.70).
DISCUSSION
On an everyday basis, our food-related decision making is guided by inferences about the taste of food, both conscious and unconscious. Despite this, the representation of gustatory mental imagery in the brain is still relatively poorly understood. In the present study, we recruited subjects to undergo fMRI scanning while imagining multiple basic tastes and subsequently viewing pictures of foods which strongly exhibited those basic taste qualities. The response to imagined tastes vs. imagined water strongly recruited the bilateral dorsal mid-insula, as well as the left oral somatosensory cortex, a pattern of activation which strongly resembles the response to actual taste vs. neutral stimulation across multiple gustatory neuroimaging studies (Avery et al., 2015, 2020, 2021; Small et al., 2010; Veldhuizen et al., 2011; Yeung et al., 2017).
The multivariate analyses of both tasks identified significant above-chance classification accuracy for imagined tastes as well as food pictures within the left dorsal mid-insular cortex, as revealed by a conjunction of multivariate searchlight analyses as well as an analysis using independent mid-insula ROIs, which also exhibit quality-specific multivariate patterns for directly experienced tastes (Avery et al., 2020). Furthermore, within the dorsal mid-insula, we were able to successfully cross-classify the inferred response to food pictures using the response to imagined basic tastes and vice-versa. A whole-brain multivariate searchlight analysis revealed that this effect was also remarkably specific to this cortical region. Additionally, we observed that the multivariate patterns for imagined tastes and food pictures were reliably similar and taste-quality specific within this region, suggesting a common neural code for the top-down representation of taste. Taken together, these results demonstrate the dorsal mid-insula’s role in the representation of taste quality, whether directly experienced, explicitly imagined, or automatically inferred when viewing pictures of food (Avery et al., 2021a).
In previous studies, the responses to other sensory modalities such as pain and visceral interoception (Simmons et al., 2013, Avery et al., 2017), as well as other food related properties such as oral texture (de Araujo and Rolls, 2004), chemisthesis (Kawakami et al., 2015), and flavor (Seubert et al., 2014) have also been identified within the dorsal mid-insula. Indeed, the representational structure of the food pictures responses might also contain some aspect of those other food properties. In fact, this seems quite likely, given that cross-task classification was significantly lower than within task classification. This suggests that, while those spatial patterns were similar enough for our machine learning models to successfully classify, there was a significant overall difference in their informational structure, which may be due to the greater sensory complexity of the depicted foods.
Multiple imaging modalities, including studies employing fMRI (Avery et al., 2015; 2017; 2020; Small et al., 2010; Veldhuizen et al., 2011; Yeung et al., 2017), EEG (Kobayakawa et al., 1999), MEG (Kobayakawa et al., 2005), intracranial EEG (Huang et al., 2021), and cortical electrode stimulation (Isnard et al., 2004) indicate that the dorsal mid-insula is the location of ‘primary’ gustatory cortex. However, the responsiveness of the dorsal mid-insula, not just to taste, but to inferred and explicitly imagined taste, as well as its responsiveness to other sensory modalities, mentioned above, suggest that this region is not simply a primary cortical sensory region for taste, but rather a multimodal association region. Indeed, ‘primary’ sensory regions across multiple sensory modalities show responsiveness to mental imagery and focused attention (Somers et al., 1999; Palmiero et al., 2009; Albers et al, 2013), as well as cross-modal responsiveness to sensory stimuli presented via other modalities (Vetter et al., 2014, Smith & Goodale, 2015). Moreover, in both rodent and non-human primate studies, putative gustatory regions of the insula are responsive to a broad variety of orosensory as well as visceral sensations (Hanamori et al., 1998, Scott & Plata-Salaman, 1999). In addition, in macaques, only a fraction of neurons are exclusively responsive to the basic taste qualities (Scott and Plata-Salaman, 1999), suggesting that across multiple mammalian species, gustatory cortex is inherently multimodal. Thus, the responsiveness of so-called ‘primary’ sensory regions to mental imagery and higher-order information coming from other sensory modalities may reflect one of the main functions of these cortical regions, that of generating internal models of stimuli from the environment, which can be later compared to incoming sensory signals, in order to predict the response to those stimuli (Friston, 2010; Barret and Simmons, 2015).
The present results suggest a lateral preference for the left dorsal mid-insula within the context of this study. The univariate and multivariate results of the Taste Imagining task clearly favored the left dorsal mid-insula, as classification of imagined tastes was only successful in the left insula. Within-task classification accuracy was also significantly greater within the left than right insula for both tasks. While our ROI analyses identified that cross-decoding was successful within the right dorsal mid-insula as well, cross-task classification accuracy was also significantly greater in the left-insula. Additionally, the region identified in the searchlight was larger within the left insula. As the taste prompts in the Taste Imagining task were presented via on-screen text, the left-sided bias observed in these results may be due to the left-hemispheric specialization for language (Gazzaniga, 1989). There is also previous evidence that the generation of mental images from memory depends on left hemisphere brain structures (Farah, 1989).
Outside of the insula, we also observed significant above-chance classification accuracy for both tasks with widespread regions of visual and ventro-temporal cortical (VTC) regions involved in interpreting object properties such as shape, color, and texture (Chao and Martin, 1999; Simmons et al., 2007; Kravitz et al., 2013; Lafer-Sousa et al., 2016). Those same regions (but not the insular cortex) were also identified within our multivariate searchlight analysis of non-food object pictures, as in our previous study (Avery et al., 2021a). Prior studies have identified the involvement of this region in the response to food images (Simmons et al., 2005), though meta-analysis of food-picture related neuroimaging studies has identified this in only a modest proportion (~40%) of prior studies (van der Laan et al., 2011), while other more recent studies have suggested that the VTC exhibits a degree of selectivity for images of food (Khosla et al., 2022). While this area clearly shows some ability to distinguish between the food images at the taste-category level, as identified in our current and prior study (Avery et al., 2021a), it seems likely that these results were driven by object-related feature differences among our sweet, sour, and salty food groups. Importantly, we did not identify the VTC in our cross-decoding analysis, suggesting that taste-quality per se was not among the properties of food images represented within this region.
Within the present study, we observed that the neural patterns for imagined tastes and food pictures were reliably similar and quality-specific within the dorsal mid-insula, suggesting that this region contains a common code for top-down representation of taste quality. Recent evidence from translational studies (Livneh et al., 2020), as well as current theories about the function of the insular cortex claim that it serves a central role in the maintenance of bodily homeostasis (de Araujo et al., 2012; Barrett and Simmons, 2015), and that sensory information from external channels, including, sight, smell, and taste, serve as predictive cues to indicate how foods and all other external stimuli will promote or disturb that homeostasis. Furthermore, clinical neuroimaging studies have highlighted food and interoception-related disfunction of the mid-insula in multiple disorders of mental health including major depression (Avery et al., 2014; Simmons et al., 2016; 2018), anorexia nervosa (Kerr et al., 2016), and addiction (Avery et al., 2017), suggesting that disrupted homeostatic processing lies at the heart of these disorders. Indeed, a recent neuroimaging meta-analyses, gathering studies from across a variety of mental health disorders, specifically point to the left mid-insula as a common locus of dysfunction (Nord et al., 2021; Paulus and Khalsa, 2021). Thus, understanding what and how information is represented in this cortical region may be crucially important for identifying the underlying neuropathology of these disorders of mental health.
METHODS
Participants
We recruited 22 healthy subjects (10 female) between the ages of 18 and 64 (Average(SD): 27.1(10.2) years). Ethics approval for this study was granted by the NIH Combined Neuroscience Institutional Review Board under protocol number 93-M-0170. The institutional review board of the National Institutes of Health approved all procedures, and written informed consent was obtained for all subjects. Participants were excluded from taking part in the study if they had any history of neurological injury, known genetic or medical disorders that may impact the results of neuroimaging, prenatal drug exposure, severely premature birth or birth trauma, current usage of psychotropic medications, or any exclusion criteria for magnetic resonance imaging (MRI).
Experimental Design
All fMRI scanning and behavioral data was collected at the NIH Clinical Center in Bethesda, MD. Participants were instructed to avoid eating or drinking anything besides water for two hours prior to their arrival at the testing center. This was done to ensure that participants were neither particularly hungry or full during the scanning session. Prior to scanning, participants rated their current levels of hunger, fullness, thirst, and tiredness on 100cm rating scales, which ranged from extremely (hungry, full, thirsty, or tired) to not at all (hungry, full, thirsty, or tired). Participant scanning sessions began with a high-resolution anatomical reference scan followed by fMRI scans, during which they performed our Taste Imagining task and our Food Pictures task. Importantly, all participants performed the Food Pictures task after the Taste Imagining task, to avoid the possibility that exposure to the food pictures might bias to imagine the tastes associated with any of the foods depicted during the Food Pictures task.
Taste Imagining fMRI Task
During the Taste Imaging task, participants viewed a word representing a specific substance (‘SUGAR’, ‘SALT’, ‘LEMON JUICE’, and ‘WATER’) presented in large black font (Figure 1) at the center of the display screen against a gray background. These specific substances were chosen, as they are each A) familiar tastes, B) clear examples of one primary taste quality (sweet, sour, salty, or neutral/tasteless), C) correspond well to the tastants used in prior studies of taste perception by our lab (Avery et al., 2020, 2021). When participants saw the words on screen, they were instructed to “imagine to the best of your ability that someone has placed a spoonful of that substance on your tongue”. Subjects viewed these words in randomized order during 8-second blocks, during which they continually imagined the tastes, separated by 6-second interstimulus-intervals (ISI), during which they saw a fixation cross at the center of the screen. Subjects performed this task during four scanning runs, each lasting 340 seconds in duration (5-minutes, 40-seconds). Each word presentation block was repeated six times per scanning run, for a total of 24 repetitions per stimulus. For MVPA analysis, each run was split into two run segments, which allowed us to use a total of eight run segments for subsequent MVPA analyses.
Food Pictures fMRI Task
The presentation of this task was nearly identical to the task reported in a previous study (Avery et al., 2021a), with some exceptions, detailed below. During this task, participants viewed images of various foods and non-food objects. Pictures were presented sequentially, with four pictures shown per 8-second presentation block. Each block consisted of four pictures of a specific sweet, sour, or salty food or of a specific type of non-food familiar objects. Pictures were presented at the center of the display screen against a gray background. Within a presentation block, pictures were presented for 1500ms, followed by a 500ms interstimulus interval (ISI), during which a fixation cross ‘+’ appeared on the screen. Another four-second ISI followed each presentation block (see Figure 1). Presentation blocks were presented in a pseudo-random order by picture category, with no item or picture category presented twice in a row.
The food types presented during this task were twelve foods selected and rated to be predominantly sweet (cake, honey, donuts, ice cream), sour (grapefruit, lemons, lemon candies, limes), or salty (potato chips, french fries, pretzels, crackers), by groups of online participants recruited through Amazon Mechanical Turk (Figure S1; details on food picture selection and online ratings were previously reported in Avery et al., 2021a). As previously reported, these foods were selected to be clearly recognizable, pleasant, and strongly characteristic of their respective taste quality (Avery et al., 2021a). Non-food objects were familiar objects - basketballs, tennis balls, fluorescent lightbulbs, incandescent lightbulbs, baseball gloves, flotation tubes, pencils, and marbles (Figure S1)- which were selected to roughly match the shape and color of the pictured foods. In total, participants of the fMRI study viewed 28 unique exemplars of each type of the 12 foods (336 total) and 14 unique exemplars of the non-food objects (112 total).
During half of the presentation blocks, one of the pictures was repeated, and participants were instructed to press a button on a handheld fiber optic response box whenever they identified a repeated picture. Blocks with repetition events were evenly distributed across picture categories (sweet, salty, sour, and object pictures), such that half the blocks of each category contained a repetition. Repetition blocks were also evenly distributed across food and object types, such that each food type was used in a repetition event four times and each object type was used twice. Two presentation blocks for each of the twelve foods (8 blocks sweet, 8 blocks sour, 8 blocks salty), and one presentation block for each of the eight objects were presented during each run of the imaging task (32 total). Each of the four imaging runs lasted for 388 seconds (6 minutes, 28 seconds). For MVPA analysis, each run was split into two run segments, which allowed us to use a total of eight run segments for subsequent MVPA analyses.
Imaging Methods
fMRI data was collected at the NIMH fMRI core facility at the NIH Clinical Center using a Siemens 7T-830/AS Magnetom scanner and a 32-channel head coil. Each echo-planar imaging (EPI) volume consisted of 58 1.2-mm axial slices (echo time (TE) = 23 ms, repetition time (TR) = 2000 ms, flip angle = 56 degrees, voxel size = 1.2 × 1.2 × 1.2 mm3). A Multi-Band factor of 2 was used to acquire data from multiple slices simultaneously. A GRAPPA factor of 2 was used for in-plane slice acceleration along with a 6/8 partial Fourier k-space sampling. Each slice was oriented in the axial plane, with an Anterior-to-Posterior phase encoding direction. Prior to task scans, a 1-minute EPI scan was acquired with the opposite phase encoding direction (Posterior-to-Anterior), which was used for correction of spatial distortion artifacts during preprocessing (see Image Preprocessing). An ultra-high resolution MP2RAGE sequence was used to provide an anatomical reference for the fMRI analysis (TE = 3.02 ms, TR = 6000 ms, flip angle = 5 degrees, voxel size = 0.70 × 0.70 × 0.70 mm).
Image Preprocessing
Most of the fMRI pre-processing was performed in AFNI (http://afni.nimh.nih.gov/afni). The FSL bet program was additionally used for skull-stripping the anatomical scans. A de-spiking interpolation algorithm (AFNI’s 3dDespike) was used to remove transient signal spikes from the EPI data, and a slice timing correction was then applied to the volumes of each EPI scan. All EPI volumes were registered to the very first EPI volume of the Taste Imagining task (the base-epi volume) using a 6-parameter (3 translations, 3 rotations) motion correction algorithm, and the motion estimates were saved for use as regressors in the subsequent statistical analyses. Within-run volume registration and registration to the base-epi volume were implemented in the same transformation step. A 3.6mm (3-voxel width) FWHM Gaussian smoothing kernel was then applied to the volume-registered EPI data. Finally, the signal intensity for each EPI volume was normalized to reflect percent signal change from each voxel’s mean intensity across the time-course. Anatomical scans were first co-registered to the base-epi volume and were then spatially normalized to Talairach space via an affine spatial transformation. Subject-level EPI statistical and decoding maps were only moved to standard space (using affline spatial transformations) after subject-level regression analyses (for univariate analyses) or decoding analyses (for multivariate analyses). All EPI data were left at the original spatial resolution (1.2×1.2×1.2mm3).
The EPI data collected during both tasks were separately analyzed at the subject-level using multiple linear regression models in AFNI’s 3dDeconvolve. For the FP task univariate analyses, the model included one regressor for each picture category (sweet, sour, salty, and objects). These regressors were constructed by convolution of a gamma-variate hemodynamic response function with a boxcar function having an 8-second width beginning at the onset of each presentation block. For the Taste Imagining task univariate analyses, the model included one 8-second block regressor for each of the 4 taste types (sugar, salty, lemon juice, and water). The regression model for both tasks also included regressors of non-interest to account for each run’s mean, linear, quadratic, and cubic signal trends, as well as the 6 normalized motion parameters (3 translations, 3 rotations) computed during the volume registration preprocessing. We additionally generated subject-level regression coefficient maps for use in the multivariate ROI and searchlight analyses. For both tasks, we generated a new subject-level regression models, which modeled each run segment (8 total; see task design above) separately, so that all conditions of both tasks would have eight beta coefficient maps for the purposes of model training and testing.
Analyses
Behavioral Analyses
Imaging Analyses - Univariate
We generated statistical contrast maps at the group level to identify brain regions that exhibited shared activation for imagined tastes vs. water and the sight of food vs. non-food pictures. For this analysis, we used the subject-level univariate beta-coefficient maps to perform separate group-level random-effects analyses, using the AFNI program 3dttest++. For the Taste Imagining task, we used the contrast, all imagined tastes (sugar, lemon juice, and salt) versus the imagined taste of water. For the Food Pictures task, we used the statistical contrast of all food pictures (sweet, sour, and salty) versus object pictures. Both contrast maps were separately whole-volume corrected for multiple comparisons using a cluster-size FWE correction (see Multiple Comparison Correction section). We then performed a conjunction of the two independent contrast maps to identify brain voxels significantly activated by both tasks.
Imaging Analyses - Multivariate
These analyses used a linear support-vector-machine (SVM) classification approach, implemented in The Decoding Toolbox (Hebart et al., 2014), to classify imagined tastes and food picture blocks based on their category labels. These SVM decoders were trained and tested separately on subject-level regression coefficients obtained from the Taste Imagining and Food Pictures tasks, using leave-one-run-segment-out cross-validation.
The whole-volume MVPA searchlight analyses (Kriegeskorte, et al. 2006) allowed us to identify the average classification accuracy within a multivoxel searchlight, defined as a sphere with a four-voxel radius centered on each voxel in the brain (251 voxels/433 mm3 total). For every subject, we performed separate searchlight analyses for both imaging tasks. The outputs of these searchlight analyses were voxel-wise maps of average pairwise classification accuracy versus chance (50%). Importantly, all classification analyses used, both at the searchlight and ROI-level, were conducted identically, using the average accuracy across pairwise tastes (sweet vs. sour, sweet vs. salty, sour vs. salty). The classification analyses of the Taste Imagine data did not involve the water condition, and the analysis of the Food Pictures data did not involve the non-food object condition. To evaluate the classification results at the group level, we warped the resulting classification maps to Talairach atlas space and applied a small amount of spatial smoothing (2.4 mm FWHM) to normalize the distribution of scores across the dataset. We then performed group-level random-effects analyses using the AFNI program 3dttest++. We applied an initial voxel-wise p-value threshold of p < 0.001 to each statistical map, and applied a cluster-size correction for multiple comparisons of p < 0.05 using an updated version of AFNI’s 3dClustsim (see Multiple Comparison Correction section for details). Through this procedure, we generated group-level classification accuracy maps for both the Food Pictures and Taste Perception tasks. We then created a binary conjunction of the two corrected classification maps, to identify shared brain regions present within both maps.
For the cross-task classification searchlight analysis, we trained the SVM decoder using the beta coefficients of the imagined tastes (sugar, lemon juice, and salt) from the Taste Imagining task and tested whether it could correctly predict the taste category (sweet, sour, and salty foods) of the beta coefficients from the Food Pictures task, and vice versa. The output values per-voxel were the average pairwise classification accuracy of the cross-classification analyses performed in both directions within a searchlight sphere centered on that voxel. As above, we performed group-level random-effects analyses using the AFNI program 3dttest++ and applied an initial voxel-wise p-value threshold of p < 0.005 to the resultant statistical map, with a cluster-size correction for multiple comparisons of p < 0.05 using an updated version of AFNI’s 3dClustsim (see Multiple Comparison Correction section for details). We also performed this cross-task decoding analysis at the region-of-interest level, as described below.
Insula ROI Analyses
We generated a set of independent regions-of-interest in the insular cortex to analyze the data from this task and compare to data from previous studies. We used a set of bilateral mid-insula ROIs, generated from our previous high-field, high-resolution study of taste perception (Avery et al., 2020). These insula ROIs showed a significant response to taste vs. tasteless in this previous study. In another previous study, we also identified that experienced tastes as well as the taste category of food pictures could be decoded within these insula ROIs (Avery et al., 2021a). These ROI analyses were performed using The Decoding Toolbox Within these ROIs, as with the searchlight analyses mentioned elsewhere in this section. Within these ROIs, we then compared the average pairwise classification accuracy (across pairwise tastes – see above) vs. chance (50%), both within and across tasks, using one-sample signed permutation tests. This procedure generates an empirical distribution of parameter averages by randomly flipping the sign of individual parameter values within a sample 10,000 times. The p-value is the proportion of the empirical distribution above the average parameter (accuracy) value. These p-values were then FDR corrected for multiple comparisons.
Pattern Similarity Analyses
We performed ROI-level multivariate pattern similarity analyses, using the ROIs described above. The goal of these analyses was to explicitly compare the degree of similarity between the spatial patterns for imagined tastes and food pictures within the mid-insula. This also allowed us to measure the specificity of these patterns, by identifying whether the patterns for imagined tastes are more similar to those of the matched food-picture category than the non-matched categories. We used the AFNI program 3dmaskdump to extract the voxel-wise run-segment-level beta coefficients for each condition of both imaging tasks from these insula ROIs. We used the R stats package to calculate all pairwise correlations between the average voxel-wise patterns for all imagined tastes (sugar, salt, lemon juice) and food picture categories (sweet, salty, and sour foods). We used the average fisher-transformed correlations across subjects to generate a group level correlation matrix (Figure 4d; see Figure S4 for individual correlation matrices). We subsequently compared the average on-diagonal correlation (i.e., correlation within the same imagined taste/picture category) to the average off-diagonal correlation (i.e., correlation between different tastes/picture categories) using a group-level pairwise t-test (Figure 4e).
Additional ROI Analyses
We performed an additional series of ROI analyses to examine the effects on classification accuracy of hemispheric laterality and classification type (within-task or between-task). We tested main effects of laterality and interactions within the left and right insula ROIs using a permutation-based ANOVA, implemented in the aovp function in the R-library lmPerm (https://cran.r-project.org/web/packages/lmPerm/lmPerm.pdf). We also performed an ANCOVA analysis using this program to examine the effect of hunger on classification accuracy in our mid-insula ROIs.
To address the question of whether reported difficulties in imagining a taste were related to neural responses in the Taste Imagine task, we performed another two supplemental analyses. One analysis examined whether reported difficulty imagining any condition of the task (sugar, salt, lemon, or water) was related to average bold response within the mid-insula. The resulting ANOVA used factors of condition (sugar, salt, lemon, or water), region (left mid-insula, right mid-insula), and difficulty (coded as a 1 or 0). Another analysis examined, via t-test, whether classification accuracy for the Taste Imagining task in the left mid-insula was different between subjects (of the 20 used in the main analysis) that reported difficulty imagining any of the tastes and those that did not.
Object Pictures Analyses
To test whether images of non-foods could be distinguished in the same areas of the brain as images of foods, we ran a set of supplemental decoding analyses at the whole-brain level. For this analysis, we modeled the imaging data from the Food Pictures task, during subject-level regression analysis, according to the individual item blocks presented (20 per run : 12 food + 8 non-food). We ran a whole-brain multivariate searchlight analysis (as described above) using the item-by-run-level beta coefficients for each the non-food items presented in this task, to determine the average accuracy for classifying these object pictures.
Correction for Multiple Comparisons
Multiple comparison correction was performed within a whole-volume temporal signal-to-noise ratio (TSNR) mask. This mask was constructed from the intersection of the EPI scan windows for all subjects, for both tasks, with a brain mask in atlas space. The mask was then subjected to a TSNR threshold, such that all remaining voxels within the mask had an average un-smoothed TSNR of 10 or greater. AFNI’s 3dFWHMx and 3dClustsim (using the -ACF option) were used to generate smoothness and cluster-size estimates within this mask using a spherical non-Gaussian spatial autocorrelation function. This process allows for more precise estimates of the cluster-sizes required to achieve family-wise-error (FWE) correction than previous estimates based on Gaussian random field theory (see Eklund et al, 2016; Cox et al., 2017). Using these smoothing parameters and this initial p-value threshold, this method has been demonstrated to produce corrected cluster size values approximately equal to those achieved through non-parametric permutation methods (Cox et al., 2017). Using this procedure, we applied a cluster-size correction of p-FWE < 0.05 to each whole-brain statistical map.
Supplementary Material
Highlights.
Imagining basic tastes evokes responses in gustatory mid-insular cortex
Quality of imagined tastes (sweet, sour, salty) can be decoded in this region
The taste of depicted foods can also be cross-decoded, by training on imagined tastes
ACKNOWLEDGEMENTS
This study was supported by the Intramural Research Program of the National Institute of Mental Health, National Institutes of Health, and it was conducted under NIH Clinical Study Protocol 93-M-0170 (ZIA MH002588). Clinical-trials.gov ID: NCT00001360.
Footnotes
Publisher's Disclaimer: This is a PDF file of an unedited manuscript that has been accepted for publication. As a service to our customers we are providing this early version of the manuscript. The manuscript will undergo copyediting, typesetting, and review of the resulting proof before it is published in its final form. Please note that during the production process errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain.
Data/Code Availability
The raw datasets and analysis code for the current study have been deposited in a public repository (https://openneuro.org/datasets/ds004312).
REFERENCES
- Albers AM, Kok P, Toni I, Dijkerman HC, & de Lange FP (2013). Shared Representations for Working Memory and Mental Imagery in Early Visual Cortex. Current Biology, 23(15), 1427–1431. [DOI] [PubMed] [Google Scholar]
- Avery JA, Drevets WC, Moseman SE, Bodurka J, Barcalow JC, & Simmons WK (2014). Major depressive disorder is associated with abnormal interoceptive activity and functional connectivity in the insula. Biological Psychiatry, 76(3), 258–266. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Avery JA, Kerr KL, Ingeholm JE, Burrows K, Bodurka J, & Simmons WK (2015). A common gustatory and interoceptive representation in the human mid-insula. Human Brain Mapping, 36(8), 2996–3006. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Avery JA, Gotts SJ, Kerr KL, Burrows K, Ingeholm JE, Bodurka J, et al. (2017). Convergent gustatory and viscerosensory processing in the human dorsal mid-insula. Human Brain Mapping, 38(4), 2150–2164. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Avery JA, Liu AG, Ingeholm JE, Riddell CD, Gotts SJ, & Martin A (2020). Taste Quality Representation in the Human Brain. The Journal of Neuroscience, 40(5), 1042–1052. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Avery JA, Liu AG, Ingeholm JE, Gotts SJ, & Martin A (2021a). Viewing images of foods evokes taste quality-specific activity in gustatory insular cortex. Proceedings of the National Academy of Sciences of the United States of America, 118(2), e2010932118. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Avery JA (2021b). Against gustotopic representation in the human brain: there is no Cartesian restaurant. Current Opinion in Physiology, 20, 23–28. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Barrett LF, & Simmons WK (2015). Interoceptive predictions in the brain. Nature Reviews Neuroscience, 16(7), 419–429. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Chao LL, & Martin A (1999). Cortical regions associated with perceiving, naming, and knowing about colors. Journal of cognitive neuroscience, 11(1), 25–35. [DOI] [PubMed] [Google Scholar]
- Chen K, Kogan JF, & Fontanini A (2020). Spatially Distributed Representation of Taste Quality in the Gustatory Insular Cortex of Behaving Mice. Current Biology, 31(2), 1–15. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Cox RW, Chen G, Glen DR, Reynolds RC, & Taylor PA (2017). FMRI clustering in AFNI: false-positive rates redux. Brain connectivity, 7(3), 152–171. [DOI] [PMC free article] [PubMed] [Google Scholar]
- De Araujo IE, & Rolls ET (2004). Representation in the Human Brain of Food Texture and Oral Fat. Journal of Neuroscience, 24(12), 3086–3093. [DOI] [PMC free article] [PubMed] [Google Scholar]
- De Araujo IE, Geha P, & Small DM (2012). Orosensory and Homeostatic Functions of the Insular Taste Cortex. Chemosensory Perception, 5(1), 64–79. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Eklund A, Nichols TE, & Knutsson H (2016). Cluster failure: Why fMRI inferences for spatial extent have inflated false-positive rates. Proceedings of the National Academy of Sciences of the United States of America, 113(28), 201602413–7905. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Farah MJ (1989). The neural basis of mental imagery. Trends in Neurosciences, 12(10), 395–399. [DOI] [PubMed] [Google Scholar]
- Fletcher ML, Ogg MC, Lu L, Ogg RJ, & Boughter JD Jr. (2017). Overlapping Representation of Primary Tastes in a Defined Region of the Gustatory Cortex. Journal of Neuroscience, 37(32), 7595–7605. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gazzaniga MS. Organization of the human brain. Science. 1989. Sep 1;245(4921):947–52. [DOI] [PubMed] [Google Scholar]
- Hanamori T, Kunitake T, Kato K, & Kannan H (1998). Responses of neurons in the insular cortex to gustatory, visceral, and nociceptive stimuli in rats. Journal of Neurophysiology, 79(5), 2535–2545. [DOI] [PubMed] [Google Scholar]
- Hebart MN, Görgen K, & Haynes J-D (2014). The Decoding Toolbox (TDT): a versatile software package for multivariate analyses of functional imaging data. Frontiers in Neuroinformatics, 8(174), 88. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Huang Y, Kakusa BW, Feng A, Gattas S, Shivacharan RS, Lee EB, et al. (2021). The insulo-opercular cortex encodes food-specific content under controlled and naturalistic conditions. Nature Communications, 1–12. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Isnard J, Guenot M, Sindou M, & Mauguière F (2004). Clinical manifestations of insular lobe seizures: a stereo-electroencephalographic study. Epilepsia, 45(9), 1079–1090. [DOI] [PubMed] [Google Scholar]
- Kawakami S, Sato H, Sasaki AT, Tanabe HC, Yoshida Y, Saito M, et al. (2015). The Brain Mechanisms Underlying the Perception of Pungent Taste of Capsaicin and the Subsequent Autonomic Responses. Frontiers in Human Neuroscience, 9(JAN2016), 720. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Khosla M, Murty NAR, & Kanwisher N (2022). A highly selective response to food in human visual cortex revealed by hypothesis-free voxel decomposition. Current Biology, 1–23. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kobayakawa T, Ogawa H, Kaneda H, Ayabe-Kanamura S, Endo H, & Saito S (1999). Spatio-temporal analysis of cortical activity evoked by gustatory stimulation in humans. Chemical Senses, 24(2), 201–209. [DOI] [PubMed] [Google Scholar]
- Kobayakawa T, Wakita M, Saito S, Gotow N, Sakai N, & Ogawa H (2005). Location of the primary gustatory area in humans and its properties, studied by magnetoencephalography. Chemical Senses, 30 Suppl 1, i226–7. [DOI] [PubMed] [Google Scholar]
- Kravitz DJ, Saleem KS, Baker CI, Ungerleider LG, & Mishkin M (2013). The ventral visual pathway: an expanded neural framework for the processing of object quality. Trends in Cognitive Sciences, 17(1), 26–49. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kriegeskorte N, Goebel R, & Bandettini P (2006). Information-based functional brain mapping. Proceedings of the National Academy of Sciences of the United States of America, 103(10), 3863–3868. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lafer-Sousa R, Conway BR, & Kanwisher NG (2016). Color-Biased Regions of the Ventral Visual Pathway Lie between Face- and Place-Selective Regions in Humans, as in Macaques. The Journal of Neuroscience, 36(5), 1682–1697. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Levitan D, Lin J-Y, Wachutka J, Mukherjee N, Nelson SB, & Katz DB (2019). Single and population coding of taste in the gustatory cortex of awake mice. Journal of Neurophysiology, 122(4), 1342–1356. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Livneh Y, Sugden AU, Madara JC, Essner RA, Flores VI, Sugden LA, et al. (2020). Estimation of Current and Future Physiological States in Insular Cortex. Neuron, 105(6), 1094–1111.e10. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Nord CL, Lawson RP, & Dalgleish T (2021). Disrupted Dorsal Mid-Insula Activation During Interoception Across Psychiatric Disorders. The American Journal of Psychiatry, 178(8), 761–770. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Oppenheimer SM, Gelb A, Girvin JP, & Hachinski VC (1992). Cardiovascular effects of human insular cortex stimulation. Neurology, 42(9), 1727–1732. [DOI] [PubMed] [Google Scholar]
- Palmiero M, Belardinelli MO, Nardo D, Sestieri C, Di Matteo R, D’Ausilio A, & Romani GL (2009). Mental imagery generation in different modalities activates sensory-motor areas. Cognitive Processing, 10 Suppl 2, S268–71. [DOI] [PubMed] [Google Scholar]
- Paulus MP, & Khalsa SS (2021). When you don’t feel right inside: Homeostatic dysregulation and the mid-insular cortex in psychiatric disorders. American Journal of Psychiatry, 178(8), 683–685. [DOI] [PubMed] [Google Scholar]
- Scott TR, & Plata-Salamán CR (1999). Taste in the monkey cortex. Physiology & Behavior, 67(4), 489–511. [DOI] [PubMed] [Google Scholar]
- Seubert J, Ohla K, Yokomukai Y, Kellermann T, & Lundström JN (2014). Superadditive opercular activation to food flavor is mediated by enhanced temporal and limbic coupling. Human Brain Mapping, 36(5), 1662–1676. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Simmons WK, Martin A, & Barsalou LW (2005). Pictures of appetizing foods activate gustatory cortices for taste and reward. Cerebral Cortex, 15(10), 1602–1608. [DOI] [PubMed] [Google Scholar]
- Simmons WK, Ramjee V, Beauchamp MS, McRae K, Martin A, & Barsalou LW (2007). A common neural substrate for perceiving and knowing about color. Neuropsychologia, 45(12), 2802–2810. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Simmons WK, Avery JA, Barcalow JC, Bodurka J, Drevets WC, & Bellgowan P (2013). Keeping the body in mind: insula functional organization and functional connectivity integrate interoceptive, exteroceptive, and emotional awareness. Human Brain Mapping, 34(11), 2944–2958. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Simmons WK, Burrows K, Avery JA, Kerr KL, Bodurka J, Savage CR, & Drevets WC (2016). Depression-Related Increases and Decreases in Appetite: Dissociable Patterns of Aberrant Activity in Reward and Interoceptive Neurocircuitry. The American Journal of Psychiatry, 173(4), 418–428. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Simmons WK, Burrows K, Avery JA, Kerr KL, Taylor A, Bodurka J, et al. (2018). Appetite changes reveal depression subgroups with distinct endocrine, metabolic, and immune states. Molecular Psychiatry, 25(7), 1457–1468. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Small DM (2010). Taste representation in the human insula. Brain Structure & Function, 214(5–6), 551–561. [DOI] [PubMed] [Google Scholar]
- Smith FW, & Goodale MA (2015). Decoding visual object categories in early somatosensory cortex. Cerebral Cortex, 25(4), 1020–1031. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Somers DC, Dale AM, Seiffert AE, & Tootell RB (1999). Functional MRI reveals spatially specific attentional modulation in human primary visual cortex. Proceedings of the National Academy of Sciences of the United States of America, 96(4), 1663–1668. [DOI] [PMC free article] [PubMed] [Google Scholar]
- van der Laan LN, de Ridder DTD, Viergever MA, & Smeets PAM (2011). The first taste is always with the eyes: a meta-analysis on the neural correlates of processing visual food cues. NeuroImage, 55(1), 296–303. [DOI] [PubMed] [Google Scholar]
- Veldhuizen MG, Albrecht J, Zelano C, Boesveldt S, Breslin P, & Lundström JN (2011). Identification of human gustatory cortex by activation likelihood estimation. Human Brain Mapping, 32(12), 2256–2266. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Vetter P, Smith FW, & Muckli L (2014). Decoding sound and imagery content in early visual cortex. Current Biology : CB, 24(11), 1256–1262. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Yeung AWK, Goto TK, Leung WK (2017) Basic taste processing recruits bilateral anteroventral and middle dorsal insulae: an activation likelihood estimation meta-analysis of fMRI studies. Brain Behav 7:e00655–712. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
The raw datasets and analysis code for the current study have been deposited in a public repository (https://openneuro.org/datasets/ds004312).
