Skip to main content
Elsevier Sponsored Documents logoLink to Elsevier Sponsored Documents
. 2016 Jan 15;125:61–73. doi: 10.1016/j.neuroimage.2015.10.022

Decoding the direction of imagined visual motion using 7 T ultra-high field fMRI

Thomas C Emmerling a,b,, Jan Zimmermann a,b, Bettina Sorger a,b, Martin A Frost a,b, Rainer Goebel a,b,c
PMCID: PMC4692515  PMID: 26481673

Abstract

There is a long-standing debate about the neurocognitive implementation of mental imagery. One form of mental imagery is the imagery of visual motion, which is of interest due to its naturalistic and dynamic character. However, so far only the mere occurrence rather than the specific content of motion imagery was shown to be detectable. In the current study, the application of multi-voxel pattern analysis to high-resolution functional data of 12 subjects acquired with ultra-high field 7 T functional magnetic resonance imaging allowed us to show that imagery of visual motion can indeed activate the earliest levels of the visual hierarchy, but the extent thereof varies highly between subjects. Our approach enabled classification not only of complex imagery, but also of its actual contents, in that the direction of imagined motion out of four options was successfully identified in two thirds of the subjects and with accuracies of up to 91.3% in individual subjects. A searchlight analysis confirmed the local origin of decodable information in striate and extra-striate cortex. These high-accuracy findings not only shed new light on a central question in vision science on the constituents of mental imagery, but also show for the first time that the specific sub-categorical content of visual motion imagery is reliably decodable from brain imaging data on a single-subject level.

Keywords: Decoding, Functional magnetic resonance imaging, Multi-voxel pattern analysis, Ultra-high field MRI, Visual mental imagery

Highlights

  • Four different directions of visual motion can be decoded during imagery at 7 T.

  • We found very high classification accuracies in single subjects.

  • High variability between activation patterns and accuracies of different subjects

Introduction

The nature and exact representation of complex mental imagery in the human brain is recognized as one of the keys to understanding creativity and intelligence (Ferguson, 1977, Kozhevnikov et al., 2013). If neuroimaging studies were able to reveal the neural correlates of mental imagery with a similar degree of accuracy as for perception, this may potentially allow for a better understanding of human creativity and intelligence since mental imagery seems to be a prerequisite for both (Miller, 1996).

One particularly naturalistic form of imagery is the imagination of visual motion. Visual motion is a crucial domain in daily life; the processing of visual motion is not only necessary to maintain a coherent percept of one's surrounding but also enables us to separate objects from each other and perceive depth. So far, research in cognitive neuroscience has mostly focused on static mental images such as orientations (Albers et al., 2013, Harrison and Tong, 2009), shapes (Stokes et al., 2009, Stokes et al., 2011), objects (Cichy et al., 2012, Lee et al., 2012, Reddy et al., 2010), and scenes (Johnson and Johnson, 2014). Investigating motion imagery on the other hand goes beyond static and towards dynamic imagery processes and could, thus, enable a better understanding of sustained forms of imagery. Due to its naturalistic character, motion imagery is close to and directly translatable to actual action implementation, for instance in human machine interfacing. While neural activation patterns during visual short-term memory of perceived visual motion have been shown recently (Emrich et al., 2013), scientific knowledge about neural correlates of motion imagery is still limited. This is due to the inherently private nature of imagery and inter-individual differences in personal strategies (Kozhevnikov et al., 2002, Kozhevnikov et al., 2005, Poltrock and Brown, 1984; also note the recently described condition of congenital aphantasia: Zeman et al., 2015) and brain areas (Kosslyn et al., 2001, Motes et al., 2008, Wraga et al., 2003) recruited during imagery. In addition, so far only the occurrence (categorical decoding), but not the specific content (sub-categorical decoding) of motion imagery has been decodable (Goebel et al., 1998) and the question of whether the actual specific content of imagined motion is decodable, is still to be answered.

When it comes to the neural correlates of imagery, numerous findings show that perception and imagery share neural circuits (for a review see Kosslyn et al., 2006). This suggests that also brain regions involved in visual motion imagery overlap with those involved in visual motion perception. The most important motion-responsive brain areas include direction-selective cells in V1 (Hubel and Wiesel, 1968, McLean and Palmer, 1989, Movshon et al., 1978), human area V3A (hV3A) (Tootell et al., 1997, Van Oostende et al., 1997), and the human motion complex (Cheng et al., 1995; hMT +; Dupont et al., 1994, Tootell et al., 1995). Different directions of perceived visual motion can be decoded from several of these areas in the visual cortex (Kamitani and Tong, 2005). More recently, the axis-of-motion selective columnar structure in the middle temporal area (MT) that is known from animal studies (Albright, 1984a, Albright et al., 1984, Diogo et al., 2003, Dubner and Zeki, 1971) was shown in humans using 7  Tesla (T) ultra-high field functional magnetic resonance imaging (fMRI; Zimmermann et al., 2011). In light of these previous studies, V1, V3A, and hMT + are the most likely candidates for the decoding of visual motion imagery.

In order to optimally decode information regarding imagery within these regions, high-field imaging and multivariate analyses seem promising. The recent developments of 7 T ultra-high field fMRI for humans has brought improvements in spatial specificity (Uğurbil et al., 2003, Uludağ et al., 2009) and increases in signal-to-noise ratio (SNR; Vaughan et al., 2001) over conventional 3 T fMRI. Additionally, with higher magnetic field strengths the influence of large draining veins on the blood-oxygen level dependent (BOLD) signal becomes smaller, while the signals from small vessels and capillaries become better detectable (Shmuel et al., 2007). This further enhances the functional specificity. Seminal studies have leveraged these benefits to explore the functional organization of the visual system at the level of cortical columns (Cheng et al., 2001, Goodyear and Menon, 2001, Menon et al., 1997, Yacoub et al., 2007, Yacoub et al., 2008). These structures are well known from animal studies (Bonhoeffer and Grinvald, 1991, Hubel and Wiesel, 1968, Mountcastle, 1957, Tanaka, 1996) and seem to be a crucial functional unit providing a key to understand computational mechanisms in early sensory areas and beyond. In addition to hardware advancements, new data analysis techniques like multi-voxel pattern analysis (MVPA; for an introduction see Pereira, Mitchell, and Botvinick, 2009) and information-based functional brain mapping (Kriegeskorte et al., 2006) enable the extraction of more spatially distributed information from fMRI data. However, there is an ongoing debate on the spatial scale of the signals that MVPA picks up. There are controversial findings on whether large-scale biases can explain decoding performances of MVPA decoding studies (Op de Beeck, 2010a) rather than signals from a fine-grained columnar organization of cortex (Swisher et al., 2010).

Together, ultra-high field fMRI and new analysis algorithms allow one to overcome the resolution boundaries given by conventional neuroimaging methods. Here we apply these methodological advances to the study of visual motion imagery.

The current study aims to decode the specific content of visual motion imagery, accounting for inter-individual differences in imaging strategies and neural activation patterns. In this way, we combine the described advantages of high-field brain imaging and multivariate analyses to decode not simply the occurrence (categorical decoding), but the specific content (sub-categorical decoding) of self-generated motion imagery. Mental imagery during the experiment was well controlled (participants were trained over multiple sessions with the imagery task) and as neutral as possible (the type of visual imagery was chosen to be abstract enough to not interfere with personal memories etc.), while still being completely self-generated (i.e., the experimental paradigm did not include any visual stimulation that could elicit bottom–up activations of the visual system or reactivation following e.g. a motion stimulus of a previous trial). Therewith we aimed to trigger low-level neural correlates of motion imagery and, thus, expected decodable information on the imagined motion direction in early visual areas (Kamitani and Tong, 2006) and MT (Goebel et al., 1998). Furthermore, in an exploratory way we attempted to link inter-individual differences in the brain activation patterns recruited during motion imagery to individual cognitive imagery styles assessed with self-report questionnaires.

Materials and methods

Subjects

15 healthy fMRI-experienced subjects (six females; age: 27.4 ± 6.3 years) with normal or corrected-to-normal vision volunteered in this study. They gave written informed consent and were paid for their participation. All procedures were conducted with approval from the local Ethical Committee of the Faculty of Psychology and Neuroscience at Maastricht University. Two subjects had to be excluded from the analyses because no significant activations could be detected based on a localizer scan (see below). One subject was excluded due to discomfort during the scanning resulting in problems to follow the experimental instructions. All subjects were students or employees at Maastricht University.

Procedure

In the experiment, subjects had to imagine dots that moved in one of four directions. Subjects were pseudo-randomly assigned to one of two groups. The first group imagined motion directions left, right, up, and down whereas the second group imagined the four diagonal motion directions. Subjects heard a cue on which specific motion direction they should imagine during the upcoming trial. These cues were auditorily presented numbers (1–4) and were associated to the four motion directions in a counterclockwise (subjects 1–9) or clockwise order (subjects 10–15). The orientation of this number-motion direction association (e.g., “1” = right vs. “1” = up) was systematically varied across subjects.

Training sessions

All subjects attended three training sessions during the week before the scanning session. In each training session subjects followed an adaptive visual imagery training (one run in the first session, two runs in the remaining sessions; see below) and completed two runs of the experimental task (see below) with shortened resting periods. Each training session lasted between 30 and 45 min.

Adaptive visual imagery training

Coherently moving white dots were presented on the screen together with an auditory cue and faded out after 4000 ms leaving only a white fixation dot. When indicated by a change of fixation color, subjects were instructed to imagine the faded stimulus as vividly as possible. They were asked to press a button as soon as they had a clear picture or “movie” of the stimulus in their mind and to hold the button down for as long as they “saw” this imagined stimulus. If subjects released the button before 8000 ms had passed the stimulus faded in again and they could try again by pressing the button, which faded the stimulus out. After successfully imagining the stimulus for more than 8000 ms the trial ended and was followed by a 5-point Likert scale (1: “No image at all, you only ‘know’ that you are thinking of the object” to 5: “Perfectly clear and vivid as normal vision”) to indicate the clearness of the imagery in the preceding trial. After the adaptive visual imagery training the subjects reported the used imagery strategies to the experimenter.

Experimental task

In the beginning and in the middle of each run, moving dots were presented as in the training task together with a white fixation dot. This was done I order to facilitate the subjects' recall of the precise stimulus configuration and the number-motion direction association. The four motion directions that had to be imagined by each group were presented for 4000 ms each together with the associated auditory cue (read-out number; from “1” to “4”). After this reference phase and a resting period of 12 s the first block of four trials started (see Fig. 1). We chose to present trials in blocks to maximize the number of trials while allowing subjects to recover from the cognitively demanding task of motion imagery during resting periods. When the fixation dot turned red and the first auditory cue was presented the subjects had to start to imagine the associated motion direction as vividly as possible. The following trials were indicated by the presentation of auditory cues and the subjects had to start to imagine the associated motion direction immediately. Trials lasted for six or eight seconds. The end of the last trial in a block (and thus the cue to stop all imagery) was indicated by the fixation dot turning white again. After a resting period of eight seconds, a 5-point Likert scale was presented. Subjects indicated the clearness of the imagery in the preceding trial block by moving an arrow via button press. After another resting period with a jittered duration of eight, ten, or twelve seconds the next block started. Subjects were instructed to fixate the fixation dot throughout the whole experiment. Each run consisted of ten blocks and the subjects completed four runs in the scanner (total duration of experimental runs approximately 40 min; 40 trials per condition).

Fig. 1.

Fig. 1

Experimental task. Visual stimulation is shown exemplarily for one block of four trials (red fixation dots) with preceding and succeeding resting blocks (white fixation dots). Durations are shown below each screenshot (possible jittered duration are given in brackets). In the table at the top the motion directions with their corresponding audio cues (that were played to the subject at the start of a trial) are shown (number-motion direction association varied between subjects). At the top right one frame of the reference stimulus is shown representing the reference phase.

Stimuli

Visual stimulation was created with PsychoPy (version 1.78.01; Peirce, 2007) and, in the scanning session, projected on a frosted screen at the top end of the scanner bed using an LCD projector (Panasonic, No PT-EZ57OEL; Newark, NJ, USA). Responses were collected through an MR compatible button box (Current Designs, 8-button response device, HHSC-2 × 4-C; Philadelphia, USA).

Vividness of Visual Imagery Questionnaire

After recruitment (before the training sessions) subjects filled in the Vividness of Visual Imagery Questionnaire (VVIQ; Marks, 1973). This self-report questionnaire measures subjective clearness and vividness of imagined objects and scenes with 16 items. The vividness ratings for each imagined item are given on a scale from 1 (“No image at all, you only ‘know’ that you are thinking of the object”) to 5 (“Perfectly clear and vivid as normal vision”). After scanning subjects filled in the VVIQ for a second time.

Object-Spatial Imagery and Verbal Questionnaire

After the scanning session, subjects were contacted to fill out the Object-Spatial Imagery and Verbal Questionnaire (OSIVQ; Blazhenkova and Kozhevnikov, 2009). The OSIVQ is a self-report questionnaire consisting of three scales for “object”, “spatial”, and “verbal” cognitive styles during mental imagery measured by 15 items each. In each item a statement is rated on a scale from 1 (“totally disagree”) to 5 (“totally agree”). We calculated the score on each scale for every subject as described in the original paper (Blazhenkova and Kozhevnikov, 2009).

Eye-Tracking

To check for eye-movements related to the different directions of imagined motion we recorded eye movements during the scanning session for four subjects (9–12) using an MR-compatible eye-tracker (Real Eye Nano; Avotec, Inc.; Stuart, FL, USA). Eye-tracking data were analyzed using custom code in MATLAB (version 2013a; The MATHWORKS Inc., Natick, MA, USA), code from the “EYE-EEG extension” toolbox (Dimigen et al., 2011; http://www2.hu-berlin.de/eyetracking-eeg) to detect saccades based on the algorithm by Engbert and Mergenthaler (2006), and code from the CircStat toolbox (Berens, 2009). To assess the statistical similarity of saccade directions across trials with different imagined directions we used a non-parametric multi-sample test for equal median directions as implemented in the CircStat toolbox under the null hypothesis that saccade directions did not differ between different directions of imagined motion. We also computed the circular–circular correlation between the direction of imagined motion and the saccade direction and tested it for significance. Furthermore, we analyzed the raw eye gaze position data during trials. After discarding data points during eye blinks (pupil aspect ratio lies outside a confidence interval of ± 2SD around the average pupil aspect ratio in each trial) we fitted a Minimum Volume Enclosing Ellipse (Moshtagh, 2005) to the XY data of each trial. Then we statistically tested the similarity of the ellipse rotations across trials and the circular–circular correlation with the different imagined directions employing the same methods we used for the saccade data (see above).

Scanning session

First, each subject completed a short 4-block practice version of the experimental task outside the scanner. At the beginning of the scanning session, we recorded an hMT + localizer scan as in Zimmermann et al. (2011) followed by a population receptive field (pRF) retinotopy mapping scan (Dumoulin and Wandell, 2008).

MRI acquisition

Images were acquired with a Siemens MAGNETOM 7 T scanner (Siemens; Erlangen, Germany) and a 32-channel head-coil (Nova Medical Inc.; Wilmington, MA, USA).

An anatomical dataset was acquired with a T1-weighted magnetization prepared rapid acquisition gradient echo (3D-MPRAGE) sequence (256 sagittal slices, matrix = 384 × 384, voxel size = 0.6 × 0.6 × 0.6 mm3). To correct for intensity inhomogeneities an additional gradient echo proton-density (GE-PD) dataset (same parameters as 3D-MPRAGE) was acquired subsequently.

High-resolution functional images were obtained using gradient echo (T2* weighted) echo-planar imaging (EPI) with the following parameters: echo time (TE) = 23 ms, repetition time (TR) = 2000 ms, generalized autocalibrating partially parallel acquisitions (GRAPPA) g-factor = 2, multi-band factor = 2, flip angle = 70°, number of slices = 54, matrix = 130 × 130, voxel size = 1.1 × 1.1 × 1.1 mm3. The field-of-view included occipital, temporal, and parietal areas but did not cover large parts of frontal cortex (see Fig. 2A). To correct for EPI distortions additional functional volumes (five volumes in the encoding direction and five volumes with a reversed encoding direction) were acquired right after the GE-PD dataset.

Fig. 2.

Fig. 2

Functional field of view (FOV) and region of interest (ROI) definitions on cortical surfaces. (A) The FOV is shown exemplarily in subject 2 for the functional images (red) from two perspectives (left: top view; right: left-hemispheric view). (B) The MT ROI and the visual motion-responsive areas ROI are exemplarily shown in the normal (left) and dilated (right) versions in subject 2.

Imaging data preprocessing

Functional and anatomical images were analyzed using BrainVoyager QX (version 2.8; Brain Innovation; Maastricht, The Netherlands), custom code in MATLAB (version 2013a; The MATHWORKS Inc.; Natick, MA, USA), and PyMVPA (version 2.3; Hanke et al., 2009). Anatomical images were corrected for bias field inhomogeneities by dividing the 3D-MPRAGE images by the GE-PD images and interpolated to a nominal voxel size of 0.55 mm isotropic to match a multiple of 2 of the resolution of the functional data. The detection of the white/gray matter boundary was conducted with the largely automatic segmentation tools of BrainVoyager QX. These tools perform a region-growing method that analyzes intensity histograms and subsequently correct topological errors in the detected borders to finally reconstruct the cortical surfaces (Goebel et al., 2006, Kriegeskorte and Goebel, 2001).

In order to correct for distortions in the echo-planar images we recorded 5 functional volumes of normal and reversed phase encoding. In these pairs of images distortions go in opposite directions and we used them to estimate the susceptibility-induced off-resonance field using a method similar to that described in Andersson et al. (2003) as implemented in FSL (Smith et al., 2004). After performing 3D rigid body motion correction of the remaining functional runs (aligning all subsequent runs to the first functional run) the estimated off-resonance field was used to correct for EPI distortions. Furthermore, functional data were high-pass filtered using a general linear model (GLM) Fourier basis set of two cycles sine/cosine per run (including linear trend removal). Functional runs were co-registered to the individual anatomical scan with an affine (9 parameter) transformation.

Region-of-interest definition

Regions of interest (ROIs) were then defined using data from the hMT + localizer scan and the pRF retinotopy projected onto an inflated surface reconstruction. Area MT was distinguished from MST in the hMT + complex (Huk et al., 2002, Zimmermann et al., 2011) by thresholding the respective contrast at p < 0.001 and restricting the resulting activation patch to 150 vertices per hemisphere (for creating the MT ROI). A second ROI was created by applying the same contrasts but thresholding at p < 0.05 and including all active patches in visual cortices (for creating the “visual motion-responsive areas” ROI) in each hemisphere. For extended versions of these ROIs the patches were dilated by a static factor (adding neighbor vertices at the patches' boundaries; 20-fold dilation for the MT ROI, 10-fold dilation for visual motion-responsive areas ROI) to include possibly contributing neighboring regions (see Fig. 2B and Table 2). Areas V1 through V4 were delineated in the polar angle map of the surface-based pRF analysis in both hemispheres. All surface patches were transformed back into volume space (from − 1 mm till + 3 mm from the gray/white matter segmentation boundary) to create the final volume ROIs.

Table 2.

Volumes of defined Regions of interest (ROIs).

Subject MT MT d ma ma d V1 V2 V3 V4
1 1026 (1.366) 8496 (11.308) 20958 (27.895) 40373 (53.736) 3343 (4.45) 4087 (5.44) 3173 (4.223) 2900 (3.86)
2 914 (1.217) 7497 (9.979) 32198 (42.856) 53566 (71.296) 4489 (5.975) 4541 (6.044) 4216 (5.611) 3621 (4.82)
3 1161 (1.545) 8404 (11.186) 28763 (38.284) 58326 (77.632) 5014 (6.674) 4315 (5.743) 4388 (5.84) 3334 (4.438)
4 1290 (1.717) 9344 (12.437) 27410 (36.483) 53552 (71.278) 2807 (3.736) 3708 (4.935) 3370 (4.485) 3076 (4.094)
5 1023 (1.362) 7584 (10.094) 41955 (55.842) 71583 (95.277) 4547 (6.052) 5049 (6.72) 4278 (5.694) 3678 (4.895)
6 986 (1.312) 8105 (10.788) 35565 (47.337) 63736 (84.833) 3808 (5.068) 3746 (4.986) 4088 (5.441) 3713 (4.942)
7 998 (1.328) 8145 (10.841) 38272 (50.94) 72269 (96.19) 5210 (6.935) 6244 (8.311) 6497 (8.648) 4535 (6.036)
8 1160 (1.544) 8486 (11.295) 23668 (31.502) 53451 (71.143) 3936 (5.239) 5032 (6.698) 4220 (5.617) 2912 (3.876)
9 990 (1.318) 8473 (11.278) 43215 (57.519) 75581 (100.598) 3596 (4.786) 3135 (4.173) 3034 (4.038) 3478 (4.629)
10 1013 (1.348) 8458 (11.258) 20816 (27.706) 42934 (57.145) 3820 (5.084) 3492 (4.648) 3016 (4.014) 2567 (3.417)
11 1003 (1.335) 7670 (10.209) 22128 (29.452) 44265 (58.917) 4226 (5.625) 3910 (5.204) 3431 (4.567) 3417 (4.548)
12 1012 (1.347) 8367 (11.136) 21256 (28.292) 52395 (69.738) 4500 (5.99) 3364 (4.477) 3163 (4.21) 3347 (4.455)
Avg 1048.0 (1.395) 8252.417 (10.984) 29683.67 (39.509) 56835.92 (75.649) 4108.0 (5.468) 4218.583 (5.615) 3906.167 (5.199) 3381.5 (4.501)
Std 98.825 (0.132) 484.973 (0.645) 8028.314 (10.686) 11298.38 (15.038) 663.526 (0.883) 846.353 (1.126) 933.923 (1.243) 483.801 (0.644)

Numbers of voxels (voxel size = 1.1 × 1.1 × 1.1 mm3) for all defined ROIs are shown for all subjects. In parentheses volumes are shown converted to (rounded) cm3. MT d: 20-fold dilated MT ROI; ma: visual motion-responsive areas ROI; ma d: 10-fold dilated visual motion-responsive areas ROI. Avg: Average (across subjects); Std: Standard deviation (across subjects).

Multi voxel pattern analyses

Each experimental run was z-scored to eliminate signal offsets and variance differences between runs. After masking the experimental data with the individually defined ROIs the data were split into training and testing datasets. We employed a leave-one-run-out splitting procedure to be able to cross-validate the classification performance. For each split the 1000 voxels (within the respective ROI) with the highest F-values in the respective training data were selected as features. The F values were computed as the standard fraction of between and within class variances (omnibus test). This was done to reduce the high number of voxels within ROIs in 7 T fMRI scans and in order to keep the number of features constant between subjects and ROIs. For each voxel and each trial we extracted the average of 6 s (3 TRs) as features. Averages were computed in a time window from 4 to 10 s after trial onset. The extracted features were then used for an one-vs-one 4-class classification (predicted classes were chosen based on the maximum number of votes in all binary classifications) using a linear support vector machine (SVM; LIBSVM implementation in PyMVPA; Chang and Lin, 2011). We repeated the whole analysis 1000 times with scrambled labels to obtain a distribution under the null hypothesis and tested the probability of the real classification accuracies against this distribution. To assess a group level statistic we tested the real classification performances of each subject against the individual average permutation classification accuracy by means of a Wilcoxon Signed-Ranks test. We chose a non-parametric test as a normal distribution of the decoding accuracies cannot be assumed. Furthermore, we computed Spearman rank-order correlations between mean classification accuracies (across all ROIs) and the OSIVQ questionnaire scores (object, spatial, and verbal scales). We tested these correlations for significance after an FDR-correction for multiple comparisons to reveal possible influences of cognitive imagery styles on the classification performance. To assess any differences between the group of subjects imagining horizontal/vertical directions and the group of subjects imagining diagonal directions we tested mean classification accuracies (across all ROIs) by means of an Mann − Whitney U test across the two groups.

Searchlight analysis

To assess the spatial distribution of brain areas involved in the mental imagery task without restriction to defined ROIs we performed a searchlight analysis (Kriegeskorte et al., 2006). A sphere with a radius of 4 voxels was moved through the cortical ribbon (so that the spheres central voxel always lay within − 1 mm to + 3 mm from the gray/white matter segmentation border) and defined a feature set of 257 voxels (4 voxel radius) that was in turn analyzed with the MVPA pipeline outlined above (without voxel pre-selection and permutation testing). The resulting classification accuracies were tested for significance by means of FDR-corrected Chi-square tests of the confusion matrix and projected onto the inflated surface reconstruction.

Furthermore, we mapped the direction selectivity of these brain areas. A preference map was computed by comparing the four t-value maps of single-direction contrasts and subsequently masked by the significant (p < 0.05) areas in the Chi-square searchlight map.

Univariate analysis

We performed a general linear model (GLM) analysis for two single subjects that showed very high MVPA decoding accuracies as significant univariate results are expected in such cases. Furthermore, significant univariate results and corresponding maps provide evidence that the MVPA classifier did not pick up on some artifactual or confounding signal rather than brain activity related to the motion imagery task. We used linear predictors for each experimental condition convolved with a standard two-gamma hemodynamic response function. We then computed contrasts for each of the directions of imagined motion against all other directions. We plotted four corresponding t-maps that were thresholded at q < 0.05 (FDR corrected) on the inflated cortical surfaces.

Results

Behavioral data

VVIQ scores for all subjects are shown in Fig. 3. Scores did not change significantly between before the training and after the scanning session (paired t-test; t(11) = 0.12; p = .903). OSIVQ scores on the three scales for all subjects are shown in Fig. 4. The strategies initially reported by subjects are shown in Table 1. In the third training session most subjects reported to not use particular strategies anymore but visualize the reference stimulus directly.

Fig. 3.

Fig. 3

Visual Imagery Questionnaire (VVIQ) scores. VVIQ scores from before the training (pre) and after the scanning session (post) are shown for all subjects (S1 through S12). The last column shows the group averages (Avg); error bars indicate one standard deviation.

Fig. 4.

Fig. 4

Object-Spatial Imagery and Verbal Questionnaire (OSIVQ) scores. OSIVQ scores for the three different scales “object”, “spatial”, and “verbal” are shown for all subjects (S1 through S12). The last column shows the group averages (Avg); error bars indicate one standard deviation.

Table 1.

Strategies used by subjects as reported during the first training session.

Subject Strategy
1 Upwards: bubbles rising in a glass; downwards: snow falling (blizzard); other two directions: city lights viewed from an airplane
2 Downwards: rain; no other specific strategy
3 Up/down: tetris game; left/right: tennis match
4 Rain/bubbles while visualizing the aperture to have a guidance for the imagery
5 Swimming tadpoles
6 No particular strategy
7 No particular strategy
8 Focus on the dot pattern; no other particular strategy
9 Upwards: bubbles rising in a glass; downwards: snow falling; left/right: buffalo herd running
10 Try to keep the almost faded out dots in mind
11 Try to keep the almost faded out dots in mind
12 Upwards: water hose; downward: rain

The training sessions improved the subjective ratings for the vividness of the imagery in most subjects (see Fig. 5). On group level, a paired t-test between behavioral data from the first and the last training session revealed a significant improvement for the ratings in the adaptive visual imagery training(t(11) = -3.64; p = .004) and the experimental task (t(11) = -3.45; p = .005). Data from the ratings given during the scanning sessions were unfortunately lost due to a software malfunction and, hence, could not be used for data analysis.

Fig. 5.

Fig. 5

Mean vividness ratings for adaptive visual imagery and experimental task in training sessions. Mean ratings for the vividness of imagery (1: “No image at all, you only ‘know’ that you are thinking of the object” to 5: “Perfectly clear and vivid as normal vision”) are shown for all subjects (S1 through S8) for the first and the last training session in the adaptive visual imagery training (left) and the experimental task (right). The last columns show the group averages (Avg); error bars indicate one standard deviation.

Multi voxel pattern analyses

Volumes of all defined ROIs are shown in Table 2. Group-level statistics revealed significant classification accuracies for all defined ROIs (MT (W = 9; p = .019), dilated MT (W = 0; p = .002), visual motion-responsive areas (W = 0; p = .002), dilated visual motion-responsive areas (W = 0; p = .002), V1 (W = 12; p = .034), V2 (W = 5; p = .008), V3 (W = 6; p = .01), and V4 (W = 3; p = .005)). The single-subject level MVPAs revealed significant classification accuracies for two subjects in the MT ROI, for two subjects in the dilated MT ROI, for five subjects in the V1, V2, V3, and V4 ROIs, for six subjects in the dilated visual motion-responsive areas, and for eight subjects in the visual motion-responsive areas ROIs, respectively (see Fig. 6). In subject 7 classification accuracies of up to 91,25% (V3 ROI; chance level at 25%) were observed. Mean classification accuracies (across all ROIs) did not differ significantly between the group of subjects imagining horizontal/vertical motion directions and the group of subjects imagining diagonal motion directions (U = 10; p = .128).

Fig. 6.

Fig. 6

Classification accuracies for imagery of visual motion in four different directions. Average classification accuracies across cross-validations for all subjects (S1–S12) are shown in eight different ROIs. Error bars show one standard deviation of the accuracies' distribution across cross-validations. The black horizontal lines represent the chance level (25%); the gray horizontal lines show the 95th percentile of each 1000-fold permutation test. Asterisks in single subjects indicate significant accuracies (p < 0.05) as assessed by 1000-fold permutation testing. Asterisks in the group average indicate significant accuracies as assessed by Wilcoxon Signed-Rank tests against the average classification accuracy of 1000-fold permutation testing. Orientation wheels at the right show the directions of motion that were imagined by the respective subjects in each row. MT d: 20-fold dilated MT ROI; ma: visual motion-responsive areas ROI; ma d: 10-fold dilated visual motion-responsive areas ROI.

Correlations with Object-Spatial Imagery and Verbal Questionnaire

No correlation between OSIVQ scores and classification accuracies was significant (rObject = .365 (p = .243); rSpatial = −.304 (p = .337); rVerbal = -.185 (p = .564); see Fig. 7).

Fig. 7.

Fig. 7

Scatter plots for Object-Spatial Imagery and Verbal Questionnaire (OSIVQ) scores and average classification accuracies (across all ROIs). Scores from the three OSIVQ scales “object” (blue), “spatial” (red), and “verbal” (green) are plotted against the average classification accuracies for each subject. The colored lines show linear regressions for each OSIVQ scale. Correlations and corresponding p-values are shown in titles of each subplot.

Searchlight analysis

Maps showing cortical areas with significant searchlight accuracies showed inter-individual differences (see Fig. 8). In subject 7 large parts of the early visual system showed significant decoding accuracies (see Fig. 9A). In subject 5 mainly V2, V3, and V4 areas going into the foveal confluence showed significant decoding accuracies (see Fig. 9B). In subject 6 areas V2L, V3L, and V4L (and beyond) showed significant decoding accuracies in the left hemisphere. In the right hemisphere area V4L showed significant decoding accuracies (see Fig. 9C). In subject 8 a superior parietal region showed significant decoding accuracies.

Fig. 8.

Fig. 8

Searchlight maps for eight subjects. Significant areas in the searchlight map (p < 0.05 FDR corrected; cluster-thresholded at 50 mm2) in the two inflated hemispheres of subjects 1, 2, 5, 6, 7, 9, 10, and 12 (there were no significant patches on the remaining subjects). Warmer colors (orange > yellow) indicate lower p-value. On the right the motion directions that had to be imagined are shown for each subject. LH: left hemisphere; RH: right hemisphere; lat.: lateral view; med.: medial view.

Fig. 9.

Fig. 9

Searchlight-thresholded univariate preference maps for three subjects. Preference maps for (A) subject 7, (B) subject 5, and (C) subject 6 calculated from univariate contrasts for each imagined motion direction are shown for those areas exhibiting significant effects in the searchlight map (p < 0.05 FDR corrected). Colors indicate preferred imagined motion directions and correspond to the direction wheel at the bottom right. Areas V1 through V4 are delineated with white lines. The MT ROI is delineated in white (white circular delineation only visible on the magnified right hemispheres of subjects 7 and 5).

Preference maps revealed individually distinct patterns. In subject 5 and 7 a difference in preference from central to peripheral parts of the visual field (following the eccentricity tuning) was observed.

Univariate analysis

For subject 5 and 7 single-subject GLM analyses showed significant activations for the different directions of imagined motion. Fig. 10 shows corresponding t-maps thresholded at q < 0.05 (FDR corrected) for the same areas on the inflated cortex as in Fig. 9A and B.

Fig. 10.

Fig. 10

Fig. 10

Univariate single-direction contrast maps for two subjects. T-maps (one direction vs all other directions) for (A) subject 7 and (B) subject 5 for each imagined motion direction are shown (q < 0.05; FDR corrected). Colors correspond to the direction wheel at the bottom right. Gray/white map colors correspond to negative t-values. Areas V1 through V4 are delineated with white lines. The MT ROI is delineated in white (white circular delineation only visible on the magnified right hemispheres of subject 7 and 5).

Eye-tracking data

No systematic eye movements associated to the different motion imagery directions were observed in subject 9 through 12 (see Fig. 11). The test for equal median directions was not significant for any subject (S9: P(3) = 1.159 p = .763; S10: P(3) = 2.018 p = .569; S11: P(3) = 0.471 p = .925; S12: P(3) = 0.352 p = .95), that is, the saccade directions did not differ significantly between different imagined motion directions. The correlation between the direction of imagined motion and the saccade direction was also not significant for any subject (S9: r = − .124 p = .37; S10: r = .008 p = .878; S11: r = .049 p = .32; S12: r = − .02 p = .666). Furthermore, the analyses of the raw eye gaze position did neither return any significant result in the test for equal median directions (S9: P(3) = 2.514 p = .473; S10: P(3) = 7.118 p = .068; S11: P(3) = 4.48 p = .214; S12: P(3) = 4.571 p = .206) nor in the test for circular–circular correlations (S9: r = .018 p = .83; S10: r = .047 p = .56; S11: r = .04 p = .626; S12: r = .046 p = .543).

Fig. 11.

Fig. 11

Eye-tracking data from (A) one subject imagining horizontal/vertical directions (subject 9) and (B) three subjects imagining diagonal direction (subjects 10, 11, and 12). For (A) and (B) histograms of saccade directions during trials are plotted for all four different directions of imagined motion (indicated by red arrows). The number of saccades per trial is shown above each polar plot.

Discussion

Using high-resolution imaging at 7 T, we were able to successfully decode directions of imagined visual motion in fMRI data recorded from the visual system. Without any visual stimulation, subjects were able to activate their visual system so specifically that neuroimaging data recorded from areas normally activated during visual perception could predict the imagined motion direction.

High sub-categorical decoding accuracies

The decoding accuracies ranged from non-significant results in four subjects to accuracies of up to 91,25% (4-class classification) in a single subject — a level rarely reached even in studies decoding perceived motion.

Localization of decodable information

Apart from the more global visual motion-responsive areas and dilated visual motion-responsive areas ROIs, the best classification accuracies were achieved in the V3 and V4 ROIs. These seem to be predominant areas to decode direction of motion during imagery. The results from the searchlight analysis complement these findings. Decoding with significant accuracies was mainly possible in the areas targeted by the ROI analyses and more pronounced in areas V3 and V4. More than that, the results from the searchlight analysis supported the local and potentially fine-grained origin of information in our study. Relatively local brain activation patterns — the searchlight sphere had a radius of 4,4 mm (4 voxels) — predicted the imagined motion direction with substantial accuracy (up to 83,75% in subject 7). These high decoding accuracies in single spheres might hint to a fine-grained columnar organization of motion-selective cortical areas. Though there is an ongoing debate on whether MVPA is able to pick up patterns of brain activation at a spatial scale beyond the recorded spatial resolution (Carlson, 2014, Op de Beeck, 2010a, Op de Beeck, 2010b) and direction preferences in early visual areas seemed to have a rather coarse pattern, the classifiers might have picked up on a more fine-grained organization of direction-selective columns because of a biased sampling (Haynes and Rees, 2006).

Decoding in hMT +

Though the information encoded in localized MT voxels gave good classification accuracies in some subjects, optimal results were only achieved when voxels from earlier visual areas were included. This is in line with previous results revealing that the decoding of the direction of perceived visual motion achieves higher accuracies when analyzing data from V1 through V4 than from hMT + (Kamitani and Tong, 2005). Moreover, recent work by Wang et al. (2014) showed that the decoding of motion direction during perception is mainly driven by an “aperture-inward” response bias producing good classification accuracies in V1, V2, and V3, but not hMT +. Although we recorded at a higher spatial resolution and decoded imagined visual motion this bias might account for the high classification accuracies in early visual areas and low classification accuracies in hMT +.

With a functional spatial resolution of 1.1 mm (iso-voxels) we are close to the resolution needed to image columnar structures. In hMT + for instance, axis-of-motion columns are estimated to have a width of 2–2.8 mm (Zimmermann et al., 2011). A columnar organization of hMT +, where neighboring columns prefer motion directions that are opposite (i.e. 180° difference) to form axis-of-motion columns, was proposed before (e.g., Born and Bradley, 2005). Such an organization, however, might have actually impaired decoding in hMT + in our study as we recorded with a spatial resolution that might just fall short of capturing single direction-of-motion columns. In order to reveal the precise organization of direction-selective regions in early visual areas and hMT +, an even higher spatial resolution has to be reached by employing, for example, spin echo sequences that seem to be even more sensitive to local activations than gradient echo sequences (De Martino et al., 2013, Uğurbil et al., 2003, Uludağ et al., 2009, Zimmermann et al., 2011).

Univariate results

When looking at the univarite statistical maps for subject 7 and 5 we observed topographic effects in mainly early and mid-level visual areas. The t-maps for each direction of imagined motion (one-vs-all) show an alignment to borders between early visual areas and to different eccentricities (foveal vs. peripheral). While there is no clear organization of this topography the significant patches overlap with the individual thresholded searchlight accuracy maps. Individual topograhpies might be influenced by the individual imagery style and strategy.

Inter-individual differences

We found no significant correlations between decoding accuracies and any questionnaire scores. Though the subscales of the OSIVQ showed differential correlations with decoding accuracies none of these correlations became significant. It would be important to explore such correlations in bigger sample sizes to get further insight into the possible reasons for inter-individual differences during imagery decoding. This could potentially explain the mixed results in other imagery decoding studies (Kaas et al., 2010, Klein et al., 2000, Knauff et al., 2000). It has to be emphasized that the sample on which our results are based is of limited size so that we cannot rule out small or medium effects that would only be observable in bigger sample sizes.

Limitations and outlook

Due to the inherently private nature of imagery, the internal “stimulation” in this experiment was hard to control, potentially inducing additional variability. However, we argue that our findings are based on visual mental imagery.

Most importantly, we decoded the direction of motion in visual areas of the cortex, thereby ruling out many confounding sources of decodable information. Furthermore, simple physiological changes like motion artifacts or changes in breathing are too unspecific to produce the effects of the observed magnitude within the complex 4-class-classification design. Cross-modal influences on the visual system could, potentially, lead to decodable information in early visual areas. Vetter et al. (2014) give an example for cross-model influences; they showed that the content of auditory stimulation can be decoded in early visual areas. In line with this result, one could argue, that the classification results in our study are driven solely by the different auditory cues preceding the imagery trials. However, the motion imagery in our study presumably was a visuospatial imagery task with a high cognitive load (subjects reported that it would need a high cognitive engagement to maintain the continuous mental image of moving dots). In the study by Vetter and colleagues, the addition of such a visuospatial imagery task to their original auditory stimulation eliminated most of the cross-modal influences. Therefore, we do not expect our results to be merely based on the auditory cue. Moreover, the achieved classification accuracies in our study are much higher than the classification accuracies reported in Vetter et al. (2014) and, thus, likely not only an effect of the auditory cue.

It could also be argued that the decoding of motion imagery might merely be an artifact of eye movements (see also Laeng and Teodorescu, 2002). When we analyzed eye-tracking data recorded during the experiment in the fMRI scanner, there were no significant associations between the imagined motion direction and the direction of saccades. Furthermore, previous work show behavioral effects of motion imagery that cannot be explained by eye movements; Winawer et al. (2010) had subjects imagine inward and simultaneously outward moving gratings inducing a motion aftereffect. In this experiment corresponding eye movement accounting for the observed effects, would have been physiologically impossible. In a future study a comparison with a closed-eyes condition would be interesting as other cognititve processes might be involved when bottom–up visual input is missing entirely.

Our results could be explained by a purely attention-based modulation of visual areas. Winawer et al. (2010) argue that top-down attention modulation would only operate on feed-forward inputs to the visual cortex. However, in their study they find equal behavioral effects of mental imagery in closed- and open-eyes conditions, where in the closed-eyes condition feed-forward input is clearly missing. Even top-down modulations would still be in line with the hypothesis that imagery tasks use neural circuitry also involved in perception tasks. Our results could be caused by an internally produced stimulus that activates early visual cortex, and an attentional modulation based on the direction of imagined motion.

Finally, neural patterns underlying mental imagery have also been shown in frontal brain regions (e.g., Ganis et al., 2004). Tailoring the fMRI sequence to optimally record from visual areas with a high spatial resolution, however, we were not able to record data from the frontal cortex.

Conclusions and implications

Our results are of remarkable significance in single subjects and show that it is possible to reliably decode the content of complex visual imagery from neuroimaging data in single trials. Sub-categorical decoding of visual mental imagery (not ‘if’ but ‘what’) is not only relevant to advance neuroscientific knowledge, but would also enable advanced brain-computer interfaces (BCI; for reviews see Kübler and Neumann, 2005, Birbaumer and Cohen, 2007, Goebel et al., 2010, Nicolas-Alonso and Gomez-Gil, 2012). In a typical BCI setup (e.g., Sorger et al., 2012), the user is asked to perform a mental imagery tasks (e.g., imagine to move the hand vs. imagine to sing a song) that corresponds to different information entities that the user wants to encode (e.g., to steer the cursor on the screen to the left vs. to the right). The BCI then classifies these different categories of mental imagery tasks based on the acquired brain activation data and executes the associated intended actions (e.g., steering the cursor to the right). However, the possibility of sub-categorical decoding would allow expressing intentions in a much more natural way; the imagery of rightward visual motion (instead of imagining singing a song) could make a cursor on the screen go to the right. The recent trend towards explicit models of representation (Naselaris and Kay, 2015) would add to a closer connection between imagery content and intention. Though the use of BCIs can be trained (Wolpaw et al., 2002), especially novice users rely on a close connection between imagery and translated action and would most probably benefit from such an advanced setup. In order to achieve high and stable single-trial classification accuracies that are necessary for BCIs, it will be interesting to employ spin echo sequences with an even higher spatial resolution in the future. On a sub-millimeter functional resolution columnar-level spatial separation of direction-of-motion features might become more explicitly exploitable.

Conflict of interest

The authors declare no competing financial interests.

Acknowledgments

This work was supported by the European Research Council (ERC) under European Union's Seventh Framework Programme FP7/2007–2013/ERC (Grant Agreement Number 269853). We would like to thank Franziska Dambacher, Felix Dücker, Tom de Graaf, Lars Hausfeld, Valentin Kemper, and Giancarlo Valente for their help.

References

  1. Albers A.M., Kok P., Toni I., Dijkerman H.C., de Lange F.P. Shared representations for working memory and mental imagery in early visual cortex. Curr. Biol. 2013;23:1427–1431. doi: 10.1016/j.cub.2013.05.065. [DOI] [PubMed] [Google Scholar]
  2. Albright T.D. Direction and orientation selectivity of neurons in visual area MT of the macaque. J. Neurophysiol. 1984;52:1106–1130. doi: 10.1152/jn.1984.52.6.1106. [DOI] [PubMed] [Google Scholar]
  3. Albright T.D., Desimone R., Gross C.G. Columnar organization of directionally selective cells in visual area MT of the macaque. J. Neurophysiol. 1984;51:16–31. doi: 10.1152/jn.1984.51.1.16. [DOI] [PubMed] [Google Scholar]
  4. Andersson J.L.R., Skare S., Ashburner J. How to correct susceptibility distortions in spin-echo echo-planar images: application to diffusion tensor imaging. Neuroimage. 2003;20:870–888. doi: 10.1016/S1053-8119(03)00336-7. [DOI] [PubMed] [Google Scholar]
  5. Berens P. CircStat: a MATLAB toolbox for circular statistics. J. Stat. Softw. 2009:31. [Google Scholar]
  6. Birbaumer N., Cohen L.G. Brain-computer interfaces: communication and restoration of movement in paralysis. J. Physiol. 2007;579:621–636. doi: 10.1113/jphysiol.2006.125633. [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Blazhenkova O., Kozhevnikov M. The new object-spatial-verbal cognitive style model: theory and measurement. Appl. Cogn. Psychol. 2009;23:638–663. [Google Scholar]
  8. Bonhoeffer T., Grinvald A. Iso-orientation domains in cat visual cortex are arranged in pinwheel-like patterns. Nature. 1991;353:429–431. doi: 10.1038/353429a0. [DOI] [PubMed] [Google Scholar]
  9. Born R.T., Bradley D.C. Structure and function of visual area MT. Annu. Rev. Neurosci. 2005;28:157–189. doi: 10.1146/annurev.neuro.26.041002.131052. [DOI] [PubMed] [Google Scholar]
  10. Carlson T.A. Orientation decoding in human visual cortex: new insights from an unbiased perspective. J. Neurosci. 2014;34:8373–8383. doi: 10.1523/JNEUROSCI.0548-14.2014. [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Chang C.-C., Lin C.-J. Libsvm. ACM Trans. Intell. Syst. Technol. 2011;2:1–27. [Google Scholar]
  12. Cheng K., Fujita H., Kanno I., Miura S., Tanaka K. Human cortical regions activated by wide-field visual motion: an H2(15)O PET study. J. Neurophysiol. 1995;74:413–427. doi: 10.1152/jn.1995.74.1.413. [DOI] [PubMed] [Google Scholar]
  13. Cheng K., Waggoner R.A., Tanaka K. Human ocular dominance columns as revealed by high-field functional magnetic resonance imaging. Neuron. 2001;32:359–374. doi: 10.1016/s0896-6273(01)00477-9. [DOI] [PubMed] [Google Scholar]
  14. Cichy R.M., Heinzle J., Haynes J.-D. Imagery and perception share cortical representations of content and location. Cereb. Cortex. 2012;22:372–380. doi: 10.1093/cercor/bhr106. [DOI] [PubMed] [Google Scholar]
  15. De Martino F., Zimmermann J., Muckli L., Uğurbil K., Yacoub E., Goebel R. Cortical depth dependent functional responses in humans at 7 T: improved specificity with 3D GRASE. PLoS One. 2013;8:e60514. doi: 10.1371/journal.pone.0060514. [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Dimigen O., Sommer W., Hohlfeld A., Jacobs A.M., Kliegl R. Coregistration of eye movements and EEG in natural reading: analyses and review. J. Exp. Psychol. Gen. 2011;140:552–572. doi: 10.1037/a0023885. [DOI] [PubMed] [Google Scholar]
  17. Diogo A.C.M., Soares J.G.M., Koulakov A., Albright T.D., Gattass R. Electrophysiological imaging of functional architecture in the cortical middle temporal visual area of Cebus apella monkey. J. Neurosci. 2003;23:3881–3898. doi: 10.1523/JNEUROSCI.23-09-03881.2003. [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Dubner R., Zeki S.M. Response properties and receptive fields of cells in an anatomically defined region of the superior temporal sulcus in the monkey. Brain Res. 1971;35:528–532. doi: 10.1016/0006-8993(71)90494-x. [DOI] [PubMed] [Google Scholar]
  19. Dumoulin S.O., Wandell B.A. Population receptive field estimates in human visual cortex. Neuroimage. 2008;39:647–660. doi: 10.1016/j.neuroimage.2007.09.034. [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Dupont P., Orban G.A., De Bruyn B., Verbruggen A., Mortelmans L. Many areas in the human brain respond to visual motion. J. Neurophysiol. 1994;72:1420–1424. doi: 10.1152/jn.1994.72.3.1420. [DOI] [PubMed] [Google Scholar]
  21. Emrich S.M., Riggall A.C., LaRocque J.J., Postle B.R. Distributed patterns of activity in sensory cortex reflect the precision of multiple items maintained in visual short-term memory. J. Neurosci. 2013;33:6516–6523. doi: 10.1523/JNEUROSCI.5732-12.2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Engbert R., Mergenthaler K. Microsaccades are triggered by low retinal image slip. Proc. Natl. Acad. Sci. U. S. A. 2006;103:7192–7197. doi: 10.1073/pnas.0509557103. [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Ferguson E.S. The mind's eye: nonverbal thought in technology. Science. 1977;197:827–836. doi: 10.1126/science.197.4306.827. [DOI] [PubMed] [Google Scholar]
  24. Ganis G., Thompson W.L., Kosslyn S.M. Brain areas underlying visual mental imagery and visual perception: an fMRI study. Brain Res. Cogn. Brain Res. 2004;20:226–241. doi: 10.1016/j.cogbrainres.2004.02.012. [DOI] [PubMed] [Google Scholar]
  25. Goebel R., Khorram-Sefat D., Muckli L., Hacker H., Singer W. The constructive nature of vision: direct evidence from functional magnetic resonance imaging studies of apparent motion and motion imagery. Eur. J. Neurosci. 1998;10:1563–1573. doi: 10.1046/j.1460-9568.1998.00181.x. [DOI] [PubMed] [Google Scholar]
  26. Goebel R., Esposito F., Formisano E. Analysis of functional image analysis contest (FIAC) data with BrainVoyager QX: from single-subject to cortically aligned group general linear model analysis and self-organizing group independent component analysis. Hum. Brain Mapp. 2006;27:392–401. doi: 10.1002/hbm.20249. [DOI] [PMC free article] [PubMed] [Google Scholar]
  27. Goebel R., Zilverstand A., Sorger B. Real-time fMRI-based brain–computer interfacing for neurofeedback therapy and compensation of lost motor functions. Imaging Med. 2010;2:407–415. [Google Scholar]
  28. Goodyear B.G., Menon R.S. Brief visual stimulation allows mapping of ocular dominance in visual cortex using fMRI. Hum. Brain Mapp. 2001;14:210–217. doi: 10.1002/hbm.1053. [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Hanke M., Halchenko Y.O., Sederberg P.B., Hanson S.J., Haxby J.V., Pollmann S. PyMVPA: a python toolbox for multivariate pattern analysis of fMRI data. Neuroinformatics. 2009;7:37–53. doi: 10.1007/s12021-008-9041-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  30. Harrison S.A., Tong F. Decoding reveals the contents of visual working memory in early visual areas. Nature. 2009;458:632–635. doi: 10.1038/nature07832. [DOI] [PMC free article] [PubMed] [Google Scholar]
  31. Haynes J.-D., Rees G. Decoding mental states from brain activity in humans. Nat. Rev. Neurosci. 2006;7:523–534. doi: 10.1038/nrn1931. [DOI] [PubMed] [Google Scholar]
  32. Hubel D., Wiesel T. Receptive fields and functional architecture of monkey striate cortex. J. Physiol. 1968;67:40–113. doi: 10.1113/jphysiol.1968.sp008455. [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Huk A.C., Dougherty R.F., Heeger D.J. Retinotopy and functional subdivision of human areas MT and MST. J. Neurosci. 2002;22:7195–7205. doi: 10.1523/JNEUROSCI.22-16-07195.2002. (doi:20026661) [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Johnson M.R., Johnson M.K. Decoding individual natural scene representations during perception and imagery. Front. Hum. Neurosci. 2014;8:59. doi: 10.3389/fnhum.2014.00059. [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Kaas A., Weigelt S., Roebroeck A., Kohler A., Muckli L. Imagery of a moving object: the role of occipital cortex and human MT/V5 + Neuroimage. 2010;49:794–804. doi: 10.1016/j.neuroimage.2009.07.055. [DOI] [PubMed] [Google Scholar]
  36. Kamitani Y., Tong F. Decoding the visual and subjective contents of the human brain. Nat. Neurosci. 2005;8:679–685. doi: 10.1038/nn1444. [DOI] [PMC free article] [PubMed] [Google Scholar]
  37. Kamitani Y., Tong F. Decoding seen and attended motion directions from activity in the human visual cortex. Curr. Biol. 2006;16:1096–1102. doi: 10.1016/j.cub.2006.04.003. [DOI] [PMC free article] [PubMed] [Google Scholar]
  38. Klein I., Paradis A.L., Poline J.B., Kosslyn S.M., Le Bihan D. Transient activity in the human calcarine cortex during visual–mental imagery: an event-related fMRI study. J. Cogn. Neurosci. 2000;12(Suppl. 2):15–23. doi: 10.1162/089892900564037. [DOI] [PubMed] [Google Scholar]
  39. Knauff M., Kassubek J., Mulack T., Greenlee M.W. Cortical activation evoked by visual mental imagery as measured by fMRI. Neuroreport. 2000;11:3957–3962. doi: 10.1097/00001756-200012180-00011. [DOI] [PubMed] [Google Scholar]
  40. Kosslyn S.M., Thompson W.L., Wraga M., Alpert N.M. Imagining rotation by endogenous versus exogenous forces: distinct neural mechanisms. Neuroreport. 2001;12:2519–2525. doi: 10.1097/00001756-200108080-00046. [DOI] [PubMed] [Google Scholar]
  41. Kosslyn S.M., Thompson W.L., Ganis G. Oxford University Press; 2006. The case for mental imagery. [Google Scholar]
  42. Kozhevnikov M., Hegarty M., Mayer R.E. Revising the visualizer–verbalizer dimension: evidence for two types of visualizers. Cogn. Instr. 2002;20:47–77. [Google Scholar]
  43. Kozhevnikov M., Kosslyn S.M., Shephard J. Spatial versus object visualizers: a new characterization of visual cognitive style. Mem. Cogn. 2005;33:710–726. doi: 10.3758/bf03195337. [DOI] [PubMed] [Google Scholar]
  44. Kozhevnikov M., Kozhevnikov M., Yu C.J., Blazhenkova O. Creativity, visualization abilities, and visual cognitive style. Br. J. Educ. Psychol. 2013;83:196–209. doi: 10.1111/bjep.12013. [DOI] [PubMed] [Google Scholar]
  45. Kriegeskorte N., Goebel R. An efficient algorithm for topologically correct segmentation of the cortical sheet in anatomical mr volumes. Neuroimage. 2001;14:329–346. doi: 10.1006/nimg.2001.0831. [DOI] [PubMed] [Google Scholar]
  46. Kriegeskorte N., Goebel R., Bandettini P. Information-based functional brain mapping. Proc. Natl. Acad. Sci. U. S. A. 2006;103:3863–3868. doi: 10.1073/pnas.0600244103. [DOI] [PMC free article] [PubMed] [Google Scholar]
  47. Kübler A., Neumann N. Brain-computer interfaces—the key for the conscious brain locked into a paralyzed body. Prog. Brain Res. 2005;150:513–525. doi: 10.1016/S0079-6123(05)50035-9. [DOI] [PubMed] [Google Scholar]
  48. Laeng B., Teodorescu D.-S. Eye scanpaths during visual imagery reenact those of perception of the same visual scene. Cogn. Sci. 2002;26:207–231. [Google Scholar]
  49. Lee S.-H., Kravitz D.J., Baker C.I. Disentangling visual imagery and perception of real-world objects. Neuroimage. 2012;59:4064–4073. doi: 10.1016/j.neuroimage.2011.10.055. [DOI] [PMC free article] [PubMed] [Google Scholar]
  50. Marks D.F. Visual imagery differences in the recall of pictures. Br. J. Psychol. 1973;64:17–24. doi: 10.1111/j.2044-8295.1973.tb01322.x. [DOI] [PubMed] [Google Scholar]
  51. McLean J., Palmer L.A. Contribution of linear spatiotemporal receptive field structure to velocity selectivity of simple cells in area 17 of cat. Vis. Res. 1989;29:675–679. doi: 10.1016/0042-6989(89)90029-1. [DOI] [PubMed] [Google Scholar]
  52. Menon R.S., Ogawa S., Strupp J.P., Uğurbil K. Ocular dominance in human V1 demonstrated by functional magnetic resonance imaging. J. Neurophysiol. 1997;77:2780–2787. doi: 10.1152/jn.1997.77.5.2780. [DOI] [PubMed] [Google Scholar]
  53. Miller A.I. Springer New York; New York, NY: 1996. Insights of genius. [Google Scholar]
  54. Moshtagh N. Minimum volume enclosing ellipsoids. Convex Optim. 2005:1–9. (doi:10.1.1.116.7691) [Google Scholar]
  55. Motes M.A., Malach R., Kozhevnikov M. Object-processing neural efficiency differentiates object from spatial visualizers. Neuroreport. 2008;19:1727–1731. doi: 10.1097/WNR.0b013e328317f3e2. [DOI] [PubMed] [Google Scholar]
  56. Mountcastle V.B. Modality and topographic properties of single neurons of cat's somatic sensory cortex. J. Neurophysiol. 1957;20:408–434. doi: 10.1152/jn.1957.20.4.408. [DOI] [PubMed] [Google Scholar]
  57. Movshon J.A., Thompson I.D., Tolhurst D.J. Spatial and temporal contrast sensitivity of neurones in areas 17 and 18 of the cat's visual cortex. J. Physiol. 1978;283:101–120. doi: 10.1113/jphysiol.1978.sp012490. [DOI] [PMC free article] [PubMed] [Google Scholar]
  58. Naselaris T., Kay K.N. Resolving ambiguities of MVPA using explicit models of representation. Trends Cogn. Sci. 2015;19:551–554. doi: 10.1016/j.tics.2015.07.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  59. Nicolas-Alonso L.F., Gomez-Gil J. Brain computer interfaces, a review. Sensors (Basel) 2012;12:1211–1279. doi: 10.3390/s120201211. [DOI] [PMC free article] [PubMed] [Google Scholar]
  60. Op de Beeck H.P. Against hyperacuity in brain reading: spatial smoothing does not hurt multivariate fMRI analyses? Neuroimage. 2010;49:1943–1948. doi: 10.1016/j.neuroimage.2009.02.047. [DOI] [PubMed] [Google Scholar]
  61. Op de Beeck H.P. Probing the mysterious underpinnings of multi-voxel fMRI analyses. Neuroimage. 2010;50:567–571. doi: 10.1016/j.neuroimage.2009.12.072. [DOI] [PubMed] [Google Scholar]
  62. Peirce J.W. PsychoPy—psychophysics software in python. J. Neurosci. Methods. 2007;162:8–13. doi: 10.1016/j.jneumeth.2006.11.017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  63. Pereira F., Mitchell T., Botvinick M. Machine learning classifiers and fMRI: a tutorial overview. Neuroimage. 2009;45:S199–S209. doi: 10.1016/j.neuroimage.2008.11.007. [DOI] [PMC free article] [PubMed] [Google Scholar]
  64. Poltrock S., Brown P. Individual differences in visual imagery and spatial ability. Intelligence. 1984;138:93–138. [Google Scholar]
  65. Reddy L., Tsuchiya N., Serre T. Reading the mind's eye: decoding category information during mental imagery. Neuroimage. 2010;50:818–825. doi: 10.1016/j.neuroimage.2009.11.084. [DOI] [PMC free article] [PubMed] [Google Scholar]
  66. Shmuel A., Yacoub E., Chaimow D., Logothetis N.K., Uğurbil K. Spatio-temporal point-spread function of fMRI signal in human gray matter at 7 Tesla. Neuroimage. 2007;35:539–552. doi: 10.1016/j.neuroimage.2006.12.030. [DOI] [PMC free article] [PubMed] [Google Scholar]
  67. Smith S.M., Jenkinson M., Woolrich M.W., Beckmann C.F., Behrens T.E.J., Johansen-Berg H., Bannister P.R., De Luca M., Drobnjak I., Flitney D.E., Niazy R.K., Saunders J., Vickers J., Zhang Y., De Stefano N., Brady J.M., Matthews P.M. Advances in functional and structural MR image analysis and implementation as FSL. Neuroimage. 2004;23(Suppl. 1):S208–S219. doi: 10.1016/j.neuroimage.2004.07.051. [DOI] [PubMed] [Google Scholar]
  68. Sorger B., Reithler J., Dahmen B., Goebel R. A real-time fMRI-based spelling device immediately enabling robust motor-independent communication. Curr. Biol. 2012;22:1333–1338. doi: 10.1016/j.cub.2012.05.022. [DOI] [PubMed] [Google Scholar]
  69. Stokes M., Thompson R., Cusack R., Duncan J. Top–down activation of shape-specific population codes in visual cortex during mental imagery. J. Neurosci. 2009;29:1565–1572. doi: 10.1523/JNEUROSCI.4657-08.2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  70. Stokes M., Saraiva A., Rohenkohl G., Nobre A.C. Imagery for shapes activates position-invariant representations in human visual cortex. Neuroimage. 2011;56:1540–1545. doi: 10.1016/j.neuroimage.2011.02.071. [DOI] [PubMed] [Google Scholar]
  71. Swisher J.D., Gatenby J.C., Gore J.C., Wolfe B.A., Moon C.-H., Kim S.-G., Tong F. Multiscale pattern analysis of orientation-selective activity in the primary visual cortex. J. Neurosci. 2010;30:325–330. doi: 10.1523/JNEUROSCI.4811-09.2010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  72. Tanaka K. Inferotemporal cortex and object vision. Annu. Rev. Neurosci. 1996;19:109–139. doi: 10.1146/annurev.ne.19.030196.000545. [DOI] [PubMed] [Google Scholar]
  73. Tootell R.B., Reppas J.B., Kwong K.K., Malach R., Born R.T., Brady T.J., Rosen B.R., Belliveau J.W. Functional analysis of human MT and related visual cortical areas using magnetic resonance imaging. J. Neurosci. 1995;15:3215–3230. doi: 10.1523/JNEUROSCI.15-04-03215.1995. [DOI] [PMC free article] [PubMed] [Google Scholar]
  74. Tootell R.B.H., Mendola J.D., Hadjikhani N.K., Ledden P.J., Liu A.K., Reppas J.B., Sereno M.I., Dale A.M. Functional analysis of V3A and related areas in human visual cortex. J. Neurosci. 1997;17:7060–7078. doi: 10.1523/JNEUROSCI.17-18-07060.1997. [DOI] [PMC free article] [PubMed] [Google Scholar]
  75. Uğurbil K., Adriany G., Andersen P., Chen W., Garwood M., Gruetter R., Henry P.-G., Kim S.-G., Lieu H., Tkac I., Vaughan T., van de Moortele P.-F., Yacoub E., Zhu X.-H. Ultrahigh field magnetic resonance imaging and spectroscopy. Magn. Reson. Imaging. 2003;21:1263–1281. doi: 10.1016/j.mri.2003.08.027. [DOI] [PubMed] [Google Scholar]
  76. Uludağ K., Müller-Bierl B., Uğurbil K. An integrative model for neuronal activity-induced signal changes for gradient and spin echo functional imaging. Neuroimage. 2009;48:150–165. doi: 10.1016/j.neuroimage.2009.05.051. [DOI] [PubMed] [Google Scholar]
  77. Van Oostende S., Sunaert S., Van Hecke P., Marchal G., Orban G.A. The kinetic occipital (KO) region in man: an fMRI study. Cereb. Cortex. 1997;7:690–701. doi: 10.1093/cercor/7.7.690. [DOI] [PubMed] [Google Scholar]
  78. Vaughan J.T., Garwood M., Collins C.M., Liu W., DelaBarre L., Adriany G., Andersen P., Merkle H., Goebel R., Smith M.B., Uğurbil K. 7 T vs. 4 T: RF power, homogeneity, and signal-to-noise comparison in head images. Magn. Reson. Med. 2001;46:24–30. doi: 10.1002/mrm.1156. [DOI] [PubMed] [Google Scholar]
  79. Vetter P., Smith F.W., Muckli L. Decoding sound and imagery content in early visual cortex. Curr. Biol. 2014;24:1256–1262. doi: 10.1016/j.cub.2014.04.020. [DOI] [PMC free article] [PubMed] [Google Scholar]
  80. Wang H.X., Merriam E.P., Freeman J., Heeger D.J. Motion direction biases and decoding in human visual cortex. J. Neurosci. 2014;34:12601–12615. doi: 10.1523/JNEUROSCI.1034-14.2014. [DOI] [PMC free article] [PubMed] [Google Scholar]
  81. Winawer J., Huk A.C., Boroditsky L. A motion aftereffect from visual imagery of motion. Cognition. 2010;114:276–284. doi: 10.1016/j.cognition.2009.09.010. [DOI] [PubMed] [Google Scholar]
  82. Wolpaw J.R., Birbaumer N., McFarland D.J., Pfurtscheller G., Vaughan T.M. Brain-computer interfaces for communication and control. Clin. Neurophysiol. 2002;113:767–791. doi: 10.1016/s1388-2457(02)00057-3. [DOI] [PubMed] [Google Scholar]
  83. Wraga M., Thompson W.L., Alpert N.M., Kosslyn S.M. Implicit transfer of motor strategies in mental rotation. Brain Cogn. 2003;52:135–143. doi: 10.1016/s0278-2626(03)00033-2. [DOI] [PubMed] [Google Scholar]
  84. Yacoub E., Shmuel A., Logothetis N.K., Uğurbil K. Robust detection of ocular dominance columns in humans using Hahn Spin Echo BOLD functional MRI at 7 Tesla. Neuroimage. 2007;37:1161–1177. doi: 10.1016/j.neuroimage.2007.05.020. [DOI] [PMC free article] [PubMed] [Google Scholar]
  85. Yacoub E., Harel N., Ugurbil K. High-field fMRI unveils orientation columns in humans. Proc. Natl. Acad. Sci. U. S. A. 2008;105:10607–10612. doi: 10.1073/pnas.0804110105. [DOI] [PMC free article] [PubMed] [Google Scholar]
  86. Zeman A., Dewar M., Della Sala S. Lives without imagery — congenital aphantasia. Cortex. 2015;3–5 doi: 10.1016/j.cortex.2015.05.019. [DOI] [PubMed] [Google Scholar]
  87. Zimmermann J., Goebel R., De Martino F., van de Moortele P.-F., Feinberg D., Adriany G., Chaimow D., Shmuel A., Uğurbil K., Yacoub E. Mapping the organization of axis of motion selective features in human area MT using high-field fMRI. PLoS One. 2011;6:e28716. doi: 10.1371/journal.pone.0028716. [DOI] [PMC free article] [PubMed] [Google Scholar]

RESOURCES