Skip to main content
Frontiers in Human Neuroscience logoLink to Frontiers in Human Neuroscience
. 2014 May 27;8:357. doi: 10.3389/fnhum.2014.00357

Empirical neuroenchantment: from reading minds to thinking critically

Sabrina S Ali 1, Michael Lifshitz 2, Amir Raz 3,4,*
PMCID: PMC4034606  PMID: 24904389

Abstract

While most experts agree on the limitations of neuroimaging, the unversed public—and indeed many a scholar—often valorizes brain imaging without heeding its shortcomings. Here we test the boundaries of this phenomenon, which we term neuroenchantment. How much are individuals ready to believe when encountering improbable information through the guise of neuroscience? We introduced participants to a crudely-built mock brain scanner, explaining that the machine would measure neural activity, analyze the data, and then infer the content of complex thoughts. Using a classic magic trick, we crafted an illusion whereby the imaging technology seemed to decipher the internal thoughts of participants. We found that most students—even undergraduates with advanced standing in neuroscience and psychology, who have been taught the shortcomings of neuroimaging—deemed such unlikely technology highly plausible. Our findings highlight the influence neuro-hype wields over critical thinking.

Keywords: neuroscience, neuroimaging, bias, judgment, persuasion, critical thinking, vividness effect, allure

Introduction

Neuroimaging is on the rise, attracting attention from both academics and the popular media. Beyond transforming the neurosciences, human brain imaging has birthed neuroenchantment—a form of sub-judicious fascination with brain science. As scholars, reporters, consumers, and the general public increasingly come to appreciate the promise of imaging technology, a tendency emerges to overestimate the present state of knowledge and inflate our actual capabilities (Racine et al., 2008; Slaby and Choudhury, 2011). While trailblazing techniques permit scientists to decode basic perceptual information from ongoing neural activity (e.g., Nishimoto et al., 2011), even the most cutting-edge tools still shy away from reading minds and unlocking complex thoughts (Haynes, 2012). Nevertheless, brain scans are beginning to find their way into courtrooms as admissible evidence despite admonitions from scientific experts (Farah et al., 2014). The present study highlights the pervasiveness of neuroenchatment and examines how neuroimaging hype often interferes with critical judgment.

Spicing up arguments with the rhetorical accoutrements of neuroscience may have meaningful effects on scientific reasoning and the believability of spurious information. A seminal study showed that displaying colorful renderings from brain scans, relative to simple bar graphs or plain text, led individuals to attribute more scientific merit to cognitive research (McCabe and Castel, 2008). Thereafter, however, some accounts questioned the import of these initial findings (Farah and Hook, 2013)—citing shortcomings in the original methodology and inability to replicate (Baker et al., 2013; Hook and Farah, 2013; Michael et al., 2013; Schweitzer et al., 2013). And yet, the quality of brain images may mediate their sway over critical reasoning: compared to tame graphical representations of the brain, images that were more three-dimensional and tangible increased the perceived quality of neuroscience information (Keehner et al., 2011).

Whether or not brain images alter perceptions of scientific arguments, situating information in a neuroscience context appears to influence critical judgment. For example, embellishing arguments with gratuitous neuroscience terms prompted non-experts to rate scientific arguments more highly compared to explanations lacking neuroscientific adornment (Weisberg et al., 2008). Moreover, an independent group recently replicated these findings in a large sample (Michael et al., 2013) and additional preliminary data seem to further confirm this effect (personal communication, Deena Weisberg). Thus, individuals from a variety of backgrounds seem to succumb to the allure of neuroscience, whereby they fail to critically isolate pertinent information and separate the wheat from the chaff.

Neuroenchantment may arise from a variety of psychological sources. Often a single compelling experience can override multiple scientific accounts, instigating faith in erroneous ideas. Social psychologists have coined this phenomenon the “vividness” effect (Frey and Eagly, 1993; Stanovich, 2012). In medicine, for example, physicians frequently refer to anecdotal testimony to justify treatment choices despite robust conflicting evidence from controlled clinical trials (Lilienfeld et al., 2003). Furthermore, individuals commonly defer to experts when assessing the validity of facts or arguments outside their immediate knowledge repertoire (Keil, 2010). Such cognitive off-loading may engender uncritical consumption of information from domains highly-gated by expertise, including neuroscience.

Whereas most critical neuroscience studies to date have examined the influence of encountering written or visualized neuroimaging findings, here we decided to focus on a different aspect. Years of experience with neuroimaging alongside many an informal discussion with fellow imagers have demonstrated to us that neuroenchantment extends beyond mere fascination with high-resolution brain graphics or neuroscience vernacular. Instead, here we focus on how personal interactions with neuroimaging technology may influence personal attitudes and beliefs. Specifically, we predicted that even a minimal ruse would derail critical thinking even among students trained in the science of imaging the living human brain. The present study examined the effects of using a sham brain scanner in a way that provides one striking first-hand experience—capitalizing on the vividness effect (Stanovich, 2012). We wanted to examine whether individuals, including those who should know better, would believe a highly dubious mind-reading procedure based on neuroimaging.

Methods

Fifty-eight undergraduate students (76% female, 24% male; age range 19–46 years, M = 22, SD = 5.26) participated in an experiment where a sham brain scanner, presuming to pass as new technology, allegedly allowed experimenters to unravel the thoughts of participants based on a measurement of neural correlates. We recruited 26 participants from an advanced undergraduate course focusing on the relative merits and shortcoming of different imaging techniques. These participants, with majors spanning psychology, neuroscience, and cognitive science, comprised the Neuro group. The professor in the course (AR) repeatedly harped on the present impossibility of mind-reading and tested this information on the final examination verifying that students internalized these points. He also spoke about his background as a mentalist—a magician who performs psychological tricks, such as mind-reading—and led class demonstrations to exemplify why the public often misinterprets these effects and takes them for genuine paranormal powers. The other 32 participants were undergraduate students unfamiliar with AR enrolled in courses unrelated to psychology, neuroscience, and cognitive science.

Upon arriving at our cognitive neuroscience laboratory for a study on “The Neural Correlates of Thought,” participants encountered a rickety mock brain scanner built from discarded medical scraps from the 1960s and adorned with an old-fashioned hair-dryer dome (see Figure 1 for images of the mock scanner, and Supplementary material for a copy of the consent form). We told participants that scientists at the Montreal Neurological Institute had developed new experimental technology to decode resting state brain activity and read the human mind. We labeled the technology Spintronics and displayed warning signs around the scanning equipment similar to those found in magnetic resonance imaging (MRI) environments. Couched as an exercise in deciphering the neural correlates of thought, we asked participants to think of four items: a two-digit number, a three-digit number, a color, and a country. Participants chose freely, wrote the items on a piece of paper for later verification, and hid their note in their pocket. Following a few minutes of pre-scan preparation (e.g., taping meaningless electrodes to the face and positioning the head carefully under the hair-dryer dome) we asked participants to ponder each item they had chosen, one at a time, in response to on-screen computerized instructions. Unbeknownst to participants, the senior author (AR) taught the undergraduate experimenter (SA) a technique used by magicians to obtain the pocketed participant information. At the end of the “scanning” process, the computer presented this information on the screen with the intimation of mind-reading. Participants completed self-report measures on a computer probing their beliefs and subjective experiences concerning the experimental process (see Table 1 for details concerning time course and specific questions). We fully debriefed all participants after completion of data collection.

Figure 1.

Figure 1

The mock neuroimaging device assembled from discarded odds and ends including a scrap salon hair dryer. Throughout the mock scan, a pre-recorded video displayed rotating three-dimensional brain slices with accompanying scanner-like audio, lending the appearance of collecting and analyzing patterns of brain activity.

Table 1.

Subjective ratings of participant beliefs concerning the neuro mind-reading paradigm (on a 7-point Likert scale ranging from 0 “not at all” to 6 “extremely”).

graphic file with name fnhum-08-00357-i0001.jpg

Participants answered all questions after the mock scan in the order listed here. *p < 0.05; **p < 0.01; ***p < 0.001.

Results

Subjective ratings indicate that despite the current infeasibility of neuroscientific mindreading and the haphazard setup of our scanner, individuals were neither skeptical nor suspicious of the paradigm (see Table 1). Seventy-six percent of participants believed in the mock-equipment (i.e., a rating of 4 or above on the believability question). Further analysis using Welche's t-tests to correct for unequal variance revealed a distinction between Neuro participants and students from other disciplinary backgrounds. Neuro participants were more skeptical [t(52.29) = 2.81, p < 0.01] and less believing of the overall paradigm [t(45) = −2.05, p < 0.05]. Nevertheless, neuroimaging knowledge hardly offered an effective shield against the perils of neuroenchantment. Approximately 65% of neuroscience-educated individuals reported believing the paradigm. Furthermore, comparative analysis of data collected before and after the revelation of the pocketed information revealed that both Neuro and Other participants were more likely to believe the scanner could potentially deduce even more complex thoughts [Overall sample: t(57) = −7.48, p < 0.001; Neuro participants: t(25) = −2.67, p < 0.05; Others: t(31) = −9.59, p < 0.001].

Discussion

Our results indicate that interacting with a brain scanner, even one as sketchy as our glorified hair-dryer apparatus, clouded critical judgment and rendered dubious facts believable. Most participants not only accepted the procedure as unfeigned but also willingly extrapolated about the potential capacities of the machine. Moreover, even individuals well-versed in the shortcomings of brain imaging succumbed to the neuroenchantment of the sham scanner—despite acquiescing that their professor for the course and principle investigator for the present study (AR) formerly performed as a professional magician. For these participants, an encounter with technology that is purportedly capable of unveiling personal higher cognition eclipsed academic knowledge concerning the implausibility of such phenomena. Thus, here we show how our experimental context can lead individuals with various levels of expertise to accept science fiction as neuro fact.

Human beings are infamously irrational when entertaining certain beliefs (Kahneman, 2011). Studies probing magical thinking show that individuals often uncritically accept anomalous events even when they do not understand the underlying causal relationships (Subbotsky, 2004). Similarly, even for our educated participants, one awe-inspiring encounter with a sketchy brain scanner overruled months of information acquired through a university course addressing the limits of neuroimaging. Participants in our study appear to have slackened their critical faculties when confronted with neuroscientific equipment in a university laboratory. Thus, our study demonstrates how encountering information in a neuroscience laboratory may cloud critical thinking, even when the experience is highly implausible and presented in a farcical manner. The present experiment, however, does not permit dissociating the persuasive influence of neuroimaging contexts from the more general effects of encountering experts in a laboratory setting. Future work would need to further unravel such nuances.

Caveats

First, although we deliberately crafted our experimental setup to elicit critical judgment by appearing absurd and ramshackle (see Figure 1), we did not directly measure whether participants perceived the visual form of the scanner as implausible; however, informal reactions from participants affirmed this impression. Second, an alternative interpretation of our findings might propose that instead of a reasoning error participants responded appropriately by deferring judgment to scientific specialists given the experimental context. After all, individuals are not necessarily wrong or naive to believe in facts or causal systems outside their expertise—especially given evidence that the system works as promised. Our experimental account, however, makes no attempt to resolve these overarching philosophical questions.

Conclusions

Speculating beyond the present data, participants in our study appeared to concurrently agree with two conflicting worldviews. The first is the scientific stance of under-promise-&-over-deliver, which construes brain imaging as an imperfect but potentially powerful technique with a strong mandate for discovery. The second is the more popular over-promise-&-under-deliver mentality, which portrays brain imaging as an omniscient strategy for unscrambling the neural correlates of thought. Fueled by popular media and lay accounts, neuroenchantment further blurs fact from fad and leads to accepting tentative suppositions as indisputable fact (Racine et al., 2008; Slaby and Choudhury, 2011). Our results suggest that familiarity and academic proficiency may prove moderately effective as safeguards against neuroenchantment; yet, even participants explicitly educated about the limits of neuroimaging succumbed to our simple trick, albeit less so than their naïve peers. We are presently investigating how education may disarm the persuasive influence of neuroimaging. We hope to report on these findings before long. Critical thinking in neuroscience will likely develop as an imperative asset to negotiate the potential pitfalls of a neuro-hype climate.

Conflict of interest statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Acknowledgments

Partial support for this research came from the Canada Research Chair program, the Canadian Institutes of Health Research, the Natural Sciences and Engineering Research Council of Canada, the Fonds de Recherche du Québec—Santé, the Vanier Canada Graduate Scholarship program, and the Volkswagen Foundation.

Supplementary material

The Supplementary Material for this article can be found online at: http://www.frontiersin.org/journal/10.3389/fnhum.2014.00357/abstract

References

  1. Baker D. A., Schweitzer N. J., Risko E. F., Ware J. M. (2013). Visual attention and the neuroimage bias. PLoS ONE 8:e74449 10.1371/journal.pone.0074449 [DOI] [PMC free article] [PubMed] [Google Scholar]
  2. Farah M. J., Hook C. J. (2013). The seductive allure of “seductive allure”. Perspect. Psychol. Sci. 8, 88–90 10.1177/1745691612469035 [DOI] [PubMed] [Google Scholar]
  3. Farah M. J., Hutchinson J. B., Phelps E. A., Wagner A. D. (2014). Functional MRI-based lie detection: scientific and societal challenges. Nat. Rev. Neurosci. 15, 123–131 10.1038/nrn3665 [DOI] [PubMed] [Google Scholar]
  4. Frey K. P., Eagly A. H. (1993). Vividness can undermine the persuasiveness of messages. J. Pers. Soc. Psychol. 65, 32–44 10.1037/0022-3514.65.1.32 [DOI] [Google Scholar]
  5. Haynes J. D. (2012). Brain reading, in I Know What You're Thinking: Brain Imaging and Mental Privacy, eds Richmond S., Rees G., Edwards S. J. L. (Oxford: Oxford University Press; ), 29–40 10.1093/acprof:oso/9780199596492.003.0003 [DOI] [Google Scholar]
  6. Hook C. J., Farah M. J. (2013). Look again: effects of brain images and mind–brain dualism on lay evaluations of research. J. Cogn. Neurosci. 25, 1397–1405 10.1162/jocn_a_00407 [DOI] [PubMed] [Google Scholar]
  7. Kahneman D. (2011). Thinking, Fast and Slow. New York, NY: Farrar, Straus, and Giroux [Google Scholar]
  8. Keehner M., Mayberry L., Fischer M. H. (2011). Different clues from different views: the role of image format in public perceptions of neuroimaging results. Psychon. Bull. Rev. 18, 422–428 10.3758/s13423-010-0048-7 [DOI] [PubMed] [Google Scholar]
  9. Keil F. C. (2010). The feasibility of folk science. Cogn. Sci. 34, 826–862 10.1111/j.1551-6709.2010.01108.x [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Lilienfeld S. O., Lynn S. J., Lohr J. M. (2003). Science and pseudoscience in clinical psychology: initial thoughts, reflections, and considerations, in Science and Pseudoscience in Clinical Psychology, eds Lilenfeld S. O., Lynn S. J., L. J. M. (New York, NY: Guilford Press; ), 1–16 [Google Scholar]
  11. McCabe D. P., Castel A. D. (2008). Seeing is believing: the effect of brain images on judgments of scientific reasoning. Cognition 107, 343–352 10.1016/j.cognition.2007.07.017 [DOI] [PubMed] [Google Scholar]
  12. Michael R. B., Newman E. J., Vuorre M., Cumming G., Garry M. (2013). On the (non) persuasive power of a brain image. Psychon. Bull. Rev. 20, 720–725 10.3758/s13423-013-0391-6 [DOI] [PubMed] [Google Scholar]
  13. Nishimoto S., Vu A. T., Naselaris T., Benjamini Y., Yu B., Gallant J. L. (2011). Reconstructing visual experiences from brain activity evoked by natural movies. Curr. Biol. 21, 1641–1646 10.1016/j.cub.2011.08.031 [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Racine E., Bell E., Illes J. (2008). Can we read minds? Ethical challenges and responsibilities in the use of neuroimaging research, in Scientific, Philosophical and Ethical Perspectives in Neuroethics, eds Gioradano J., Gorijn B. (Cambridge: Cambridge University Press; ), 246–270 [Google Scholar]
  15. Schweitzer N. J., Baker D. A., Risko E. F. (2013). Fooled by the brain: re-examining the influence of neuroimages. Cognition 129, 501–511 10.1016/j.cognition.2013.08.009 [DOI] [PubMed] [Google Scholar]
  16. Slaby J., Choudhury S. (2011). Proposal for a critical neuroscience, in Critical Neuroscience: A Handbook of the Social and Cultural Contexts of Neuroscience, eds Choudhury S., Slaby J. (Oxford: Blackwell Publishing; ), 29–52 [Google Scholar]
  17. Stanovich K. E. (2012). How to Think Straight About Psychology, 10th Edn. New York, NY: Pearson [Google Scholar]
  18. Subbotsky E. (2004). Magical thinking in judgments of causation: Can anomalous phenomena affect ontological causal beliefs in children and adults? Br. J. Dev. Psychol. 22, 123–152 10.1348/026151004772901140 [DOI] [Google Scholar]
  19. Weisberg D. S., Keil F. C., Goodstein J., Rawson E., Gray J. R. (2008). The seductive allure of neuroscience explanations. J. Cogn. Neurosci. 20, 470–477 10.1162/jocn.2008.20040 [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials


Articles from Frontiers in Human Neuroscience are provided here courtesy of Frontiers Media SA

RESOURCES