Abstract
On social media and in everyday life, people are often exposed to misinformation. Decades of research have shown that exposure to misinformation can have significant impacts on people's thoughts, actions, and memories. During global pandemics like COVID‐19, people are likely exposed to heightened quantities of misinformation as they search for and are exposed to copious amounts of information about the disease and its effects. This media environment, with an abundance of both accurate and inaccurate information, is often called an “infodemic.” In the current essay, we discuss the consequences of exposure to misinformation during this infodemic, particularly in the domain of memory. We review existing research demonstrating how inaccurate, postevent information impacts a person's memory for a previously witnessed event. We discuss various factors that strengthen the impact of misinformation, including repetition and whether the misinformation is consistent with people's pre‐existing attitudes or beliefs. We conclude by describing how social media companies and individual users can help prevent the spread of misinformation and the ways in which cognitive science research can inform these approaches.
Keywords: continued influence effect, COVID‐19, infodemic, memory, misinformation, misinformation effect, misinformation prevention, pandemic, social media, warnings
1. INTRODUCTION
As the world copes with the ongoing crisis of SARS‐CoV‐2 spreading around the world, citizens are quite naturally consuming copious quantities of information about the disease's lethality, about preventing infection, and more. At the same time, concerns continue about people's exposure to misinformation (i.e., misleading or inaccurate information) about the virus from news outlets, social media, and through person‐to‐person interaction. Exposure to misinformation can have serious consequences. Several 5G cell phone towers in Europe were destroyed after misinformation circulated that these towers could spread the coronavirus (Ahmed, Vidal‐Alaball, Downing, & Lopez Segui, 2020). The World Health Organization (2020) has described the current media climate—with its abundance of both accurate and inaccurate information—an “infodemic.” This creates a unique circumstance. During the pandemic, people naturally search for accurate, trustworthy information but due to the infodemic this information may be obscured as it is intermingled with misinformation.
Indeed this concept of an infodemic is not novel to the COIVD‐19 pandemic. The concept of “infodemiology” was first discussed by Gunther Eysenbach in 2002. He originally defined an infodemic as “an excessive amount of unfiltered information concerning a problem such that the solution is made more difficult” (Eysenbach, 2009, p. 1). The ongoing discussion of infodemics is likely heightened by the profusion of information people can access online.
Particularly in the domain of health‐related behavior, a great deal of information‐seeking occurs online. In 2012, over 70% of Americans reported searching for health information online, making it one of the most popular online activities behind checking email and researching a product for purchase (Fox & Duggan, 2013; Fox & Fallows, 2003). During these searches, people are likely to encounter some form of misinformation, and exposure to misinformation can have serious negative consequences (Southwell et al., 2019). Nearly two‐thirds of Americans report that fake news or misinformation has left them confused about basic facts (Barthel, Mitchell, & Holcomb, 2016). More Americans report viewing made‐up news as a bigger problem in the country than other important issues such as illegal immigration and violent crime (Mitchell, Gottfried, Stocking, Walker, & Fedeli, 2019).
Concerns about the prevalence and impact of misinformation mushroomed after the 2016 U.S. presidential election. Indeed, in the past several years, terms like fake news, post‐truth, and misinformation have been featured as the “word of the year” in the Collins Dictionary, the Oxford Dictionary, and Dictionary.com (Brashier & Schacter, 2020). Correspondingly, there has been an increase in empirical research about the effects of misinformation on people's thoughts, beliefs, and behaviors (Wang, McKee, Torbica, & Stuckler, 2019).
Although the study of the presence and consequences of health‐related misinformation on social media is more recent, cognitive scientists have been studying misinformation in a different context for nearly 50 years. Specifically, they have examined how inaccurate, postevent misinformation affects people's memories (e.g., see work by scholars including Stephan Lewandowsky, Henry Otgaar, or reviews by Brainerd & Reyna, 2005; Loftus, 2005). This literature on the effect of misinformation on memory may help provide insight into the spread and effects of misinformation online during the COVID‐19 pandemic. To further this aim, we discuss here the varying definitions of misinformation, research from cognitive scientists about whether and when people are susceptible to misinformation, and possible technology‐based solutions to prevent misinformation spread.
2. DEFINING AND STUDYING MISINFORMATION
Despite the increased relevance of misinformation in both everyday language and in the research literature, there is no one, universally used, definition of misinformation. Misinformation can generally be defined as information that turns out to be inaccurate (Cook & Lewandowsky, 2011) or “information that is contrary to the epistemic consensus of the scientific community regarding a phenomenon” (Swire‐Thompson & Lazer, 2020, p. 434). Some definitions of misinformation focus on intentionality. If inaccurate information is distributed unintentionally or without manipulative intent, then it would be labeled as misinformation (United Nations Educational Scientific and Cultural Organization, 2018). On the other hand, if it were distributed knowingly or deliberately, it would be labeled as disinformation (Swire‐Thompson & Lazer, 2020). Determining intentionality can often be difficult so we thus consider the more encompassing definition of misinformation as information that is inaccurate regardless of intentionality (Swire‐Thompson & Lazer, 2020; Wang et al., 2019).
In the context of COVID‐19, imagine a person who reads a news story on social media that suggests that hand washing does not help prevent the spread of coronavirus. This information would be labeled as misinformation as it contradicts the generally accepted consensus of the scientific community and would be inaccurate. And there may be real‐world consequences to exposure to this misinformation. After reading this message, people's beliefs about the effectiveness of handwashing may change and their knowledge about coronavirus prevention may become more inaccurate. Yet the study of misinformation is not limited only to how inaccurate information can affect a person's belief or knowledge about an issue. Misinformation can also affect a person's memory.
Cognitive scientists have been studying the effects of misinformation on memory for nearly a half‐century: studying how exposure to it can affect a person's memory for a previously witnessed event. This ever‐growing body of cognitive research on misinformation shows what happens when people experience some event and are later exposed to misleading information about that event. In a typical study of this kind of misinformation, people see some event (e.g., a simulated crime or accident), and then are deliberately exposed to misinformation about what they saw. Sometime after that, they are tested on their event memory, and many people will incorporate elements from the misleading material into their memory for the original event. So people might have originally seen a video of a mock crime in which a thief steals a woman's wallet and hides it in his jacket pocket. But later, after being exposed to misinformation, they remember they saw the thief steal a cell phone, rather than a wallet. This phenomenon now has a scientific name: “the misinformation effect” (see Loftus, 2005 for a review). Briefly put, exposing people to misinformation can negatively affect their memory for past material.
Research on the misinformation effect may illuminate current discussions of the coronavirus. Oftentimes, when misinformation is discussed among public policy experts, journalists, and the general public, the focus is on the effects of misinformation on people's knowledge and beliefs about the causes, spread, and symptoms of COVID‐19. Findings on the misinformation effect show there is another aspect of misinformation to consider: the effect of misinformation on memory. That is, reading an article that contains misinformation about the effectiveness of masks may not only affect people's general knowledge about the value of masks, but it can also potentially change people's memory of how comfortable they were wearing a mask or how safe they felt doing so. If people remember a more negative experience wearing masks, this may impact their likelihood of doing so in the future. There are other valuable lessons to be learned from the extensive body of work on the misinformation effect that can guide the public and policymakers in their efforts to protect people during the pandemic.
3. LESSONS FROM RESEARCH ON THE EFFECTS OF MISINFORMATION ON MEMORY
Decades of research about the power of misinformation on memory have shown that virtually all of us are susceptible to misinformation. Even people with extraordinary memory abilities, who can recall accurate details of their lives all the way back to childhood, still remain susceptible to the misinformation effect (Patihis et al., 2013). There are however some known individual differences and circumstances in which misinformation susceptibility is reduced or enhanced. One of these is age. The misinformation effect occurs more strongly in young children and older adults than young adults (Sutherland & Hayne, 2001; Wylie et al., 2014).
On social media, older adults are particularly likely to share fake news articles compared to younger age groups (Guess, Nagler, & Tucker, 2019). This may occur as older adults are less digitally literate than younger adults and so may struggle more in discerning misinformation (Brashier & Schacter, 2020). The social climate in which older adults use social media also differs from other age groups. Older adults often use social media for socialization rather than information‐gathering and so information accuracy may not be a salient goal (Brashier & Schacter, 2020).
Memory research has also shown that people are more susceptible to misinformation when it fits with their preexisting attitudes or beliefs. In one study, participants saw negative, doctored photographs of Democratic and Republican politicians (Frenda, Knowles, Saletan, & Loftus, 2013). Political affiliation significantly predicted whether people reported remembering these fictional events. Participants who identified as politically conservative were more likely to remember a fictional image of a liberal politician engaging in a negative action (e.g., President Obama shaking hands with the Iranian president), while those who self‐reported as politically liberal were more likely to remember a false image of a conservative politician engaging in a negative action (e.g., President Bush vacationing with a baseball star during Hurricane Katrina). Thus demonstrating that misinformation is particularly powerful when it aligns with a person's preexisting beliefs.
Various aspects of the misinformation itself can affect the likelihood of its impact. Repetition is one influential factor in the persuasiveness of misinformation. Repeated exposure to information makes that information feel more familiar and thus more accurate (Foster, Huthwaite, Yesberg, Garry, & Loftus, 2012; Lewandowsky, Ecker, Seifert, Schwarz, & Cook, 2012). Even simply having people repeat a false rumor to themselves increases the sway of the rumor (Berinsky, 2017). Repetitions of misinformation are particularly pernicious as people are more likely to feel comfortable sharing misinformation they have been exposed to several times (Effron & Raj, 2020). Moreover, misinformation that is repeated is more difficult to correct for (Walter & Tukachinsky, 2020).
4. PREVENTING THE SPREAD OF MISINFORMATION
Importantly, this body of research about when and for whom misinformation is most influential has helped in developing ways to protect against its distorting effects (Lewandowsky et al., 2012). Unfortunately, effectively retracting misinformation can be quite difficult. Once a person has been exposed to misinformation, it can be extremely challenging to eliminate its effects fully and accurately. Indeed, even when misinformation is withdrawn, it can still continue to influence people in a phenomenon known as the continued influence effect (Ecker, Lewandowsky, & Tang, 2010). Thus, approaches to combatting misinformation that focus on preventing its spread are likely to be more effective than those that attempt to debunk misinformation after people have been exposed to it.
One commonly proposed solution is warnings. That is, informing people either before or after reading misinformation that some of the information they read may be inaccurate. Many social media companies use warnings to prevent the spread of misinformation. On Twitter, if a tweet contains disputed coronavirus information—defined as “statements or assertions in which the accuracy, truthfulness, or credibility of the claim is contested or unknown”—that has a high potential for harm, then a warning is applied to that content (Roth & Pickles, 2020). For instance, a text box may appear over the tweet informing readers that “some or all of the content shared in this Tweet conflicts with guidance from public health experts regarding COIVD‐19.”
Research has shown warnings can be effective in reducing the misinformation effect if people are explicitly warned that the information they are about to read may be inaccurate (Lewandowsky et al., 2012). Generally, warnings are more effective when they are delivered prior to people being exposed to misleading information rather than after the fact. However, postwarnings can be effective especially when they not only inform people of the presence of misinformation but also provide an explanation about why the misinformation was present (Blank & Launay, 2014). Despite this optimistic news, there may be some potential downsides of warnings. As warnings become more common on social media platforms, users might develop an expectation that stories containing misinformation are detected and labeled. Thus, stories containing misinformation that do not contain a warning may actually be considered more accurate (Pennycook, Bear, Collins, & Rand, 2020).
In addition to warnings, Facebook uses a feature to address misinformation called “related articles.” Instead of providing a warning that misinformation may be present in a news story, the related articles feature suggests additional, related reading on a person's News Feed before they click on a link to an article (Smith, Jackson, & Raj, 2017). Initial evidence suggests related articles that correct for misinformation in an initial posting do reduce reliance on misinformation (Bode & Vraga, 2015). The related articles feature can also be implemented with accurate information (i.e., providing related articles that confirm the information in an initial post), thus avoiding potential negative connotations of fact‐checking (Smith et al., 2017).
Beyond warnings, simply providing additional messaging may help prevent the spread of misinformation. Recently, Facebook implemented a feature providing additional context to articles before a user shares them (Hegeman, 2020). For instance, if a person were to read and attempt to share an article published many years ago, Facebook would provide a pop‐up message informing the reader that the article was several years old. The reader can then acknowledge this message and choose to share the article or not. This kind of system could be expanded. Social media algorithms could be modified to provide additional messages when a person tries to share misleading content. This notification could even be placed when a person opens an article rather than only when they choose to share it. For instance, messaging when a user clicks on a misleading article they have read in the past to reduce misinformation repetition.
In addition to system‐level changes implemented by social media companies, individuals can also engage in actions that reduce the spread of misinformation throughout their community. Recent research has shown that one reason people may inadvertently share misinformation on social media is because accuracy is not a salient goal (Pennycook, McPhetres, Zhang, Lu, & Rand, 2020). Thus, actions that cause people to pause and reflect before sharing can help reduce the likelihood of misinformation spread (Fazio, 2020). Social media and other online news sources can evoke this reflection process by asking participants to acknowledge messages regarding the content they wish to share. But individual users can also develop these habits on their own. The goal would be to slow down the information superhighway, turning it into a two‐lane dirt road, for a good cause. The act of pausing and reflecting before sharing information online can specifically reduce the likelihood that misinformation is shared (Bago, Rand, & Pennycook, 2020; Fazio, 2020).
5. FINAL REMARKS
In many ways, people's everyday lives during this pandemic mirror a misinformation experiment. Even with technological advances that attempt to stem the flow of misinformation, it abounds on social media and in day‐to‐day life. This problem may only grow in the future with more sophisticated types of misinformation including doctored photographs and videos. Detecting this type of doctored information is particularly difficult and thus it may be especially impactful. Decades of research have shown that misinformation, particularly health‐related misinformation, can affect people's lives in a myriad of consequential ways. Misinformation can impact beliefs about a disease's impact, effective preventive behaviors one can take, and even people's memories about their own past experiences. Thus, combating infodemics, now and in the future, is probably best accomplished with an interdisciplinary approach bringing together research and expertise from fields like technology, journalism, public policy, and cognitive science.
Biographies
Rachel Leigh Greenspan is an Assistant Professor in the Department of Criminal Justice and Legal Studies at the University of Mississippi. She received her PhD in Psychology and Social Behavior from the University of California, Irvine and completed a postdoctoral fellowship at the Quattrone Center for the Fair Administration of Justice at the University of Pennsylvania Carey School of Law. Her research interests focus on the intersection of psychology and the law exploring the development and downstream consequences of memory errors, particularly in the criminal justice system.
Elizabeth Loftus is a Distinguished Professor at the University of California—Irvine. She holds faculty positions in the Department of Psychological Science, the Department of Criminology, Law & Society, and the School of Law. She received her PhD in Psychology from Stanford University. Since then, she has published 22 books and over 600 scientific articles. Loftus's research has focused on the malleability of human memory. She has been recognized for her research with seven honorary doctorates and election to numerous prestigious societies, including the National Academy of Sciences. She is the past president of the Association for Psychological Science, the Western Psychological Association, and the American Psychology‐Law Society.
Greenspan RL, Loftus EF. Pandemics and infodemics: Research on the effects of misinformation on memory. Hum Behav & Emerg Tech. 2021;3:8–12. 10.1002/hbe2.228
DATA AVAILABILITY STATEMENT
No data were created or analyzed for the current essay.
REFERENCES
- Ahmed, W. , Vidal‐Alaball, J. , Downing, J. , & Lopez Segui, F. (2020). COVID‐19 and the 5G conspiracy theory: Social network analysis of twitter data. Journal of Medical Internet Research, 22(5), e19458. 10.2196/19458 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bago, B. , Rand, D. G. , & Pennycook, G. (2020). Fake news, fast and slow: Deliberation reduces belief in false (but not true) news headlines. Journal of Experimental Psychology: General, 149(8), 1608–1613. 10.1037/xge0000729 [DOI] [PubMed] [Google Scholar]
- Barthel, M. , Mitchell, A. , & Holcomb, J. (2016). Many Americans believe fake news is sowing confusion. Pew Research Center . Retrieved from https://www.journalism.org/2016/12/15/many-americans-believe-fake-news-is-sowing-confusion/
- Berinsky, A. J. (2017). Rumors and health care reform: Experiments in political misinformation. British Journal of Political Science, 47(2), 241–262. 10.1017/S0007123415000186 [DOI] [Google Scholar]
- Blank, H. , & Launay, C. (2014). How to protect eyewitness memory against the misinformation effect: A meta‐analysis of post‐warning studies. Journal of Applied Research in Memory and Cognition, 3(2), 77–88. 10.1016/j.jarmac.2014.03.005 [DOI] [Google Scholar]
- Bode, L. , & Vraga, E. K. (2015). In related news, that was wrong: The correction of misinformation through related stories functionality in social media. Journal of Communication, 65(4), 619–638. 10.1111/jcom.12166 [DOI] [Google Scholar]
- Brainerd, C. J. , & Reyna, V. F. . (2005). The science of false memory 10.1093/acprof:oso/9780195154054.001.0001 [DOI]
- Brashier, N. M. , & Schacter, D. L. (2020). Aging in an era of fake news. Current Directions in Psychological Science, 29(3), 316–323. 10.1177/0963721420915872 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Cook, J. , & Lewandowsky, S. (2011). The debunking handbook . Retrieved from http://sks.to/debunk
- Ecker, U. K. H. , Lewandowsky, S. , & Tang, D. T. W. (2010). Explicit warnings reduce but do not eliminate the continued influence of misinformation. Memory & Cognition, 38(8), 1087–1100. 10.3758/MC.38.8.1087 [DOI] [PubMed] [Google Scholar]
- Effron, D. A. , & Raj, M. (2020). Misinformation and morality: Encountering fake‐news headlines makes them seem less unethical to publish and share. Psychological Science, 31(1), 75–87. 10.1177/0956797619887896 [DOI] [PubMed] [Google Scholar]
- Eysenbach, G. (2002). Infodemiology: The epidemiology of (mis)information. The American Journal of Medicine, 113(9), 763–765. 10.1016/S0002-9343(02)01473-0 [DOI] [PubMed] [Google Scholar]
- Eysenbach, G. (2009). Infodemiology and infoveillance: Framework for an emerging set of public health informatics methods to analyze search, communication and publication behavior on the internet. Journal of Medical Internet Research, 11(1), e11. 10.2196/jmir.1157 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Fazio, L. (2020). Pausing to consider why a headline is true or false can help reduce the sharing of false news. Harvard Kennedy School Misinformation Review, 1(2), 1–8. 10.37016/mr-2020-009 [DOI] [Google Scholar]
- Foster, J. L. , Huthwaite, T. , Yesberg, J. A. , Garry, M. , & Loftus, E. F. (2012). Repetition, not number of sources, increases both susceptibility to misinformation and confidence in the accuracy of eyewitnesses. Acta Psychologica, 139(2), 320–326. 10.1016/j.actpsy.2011.12.004 [DOI] [PubMed] [Google Scholar]
- Fox, S. , & Duggan, M. (2013). Health online 2013. Pew Internet & American Life Project . Retrieved from https://www.pewinternet.org/wp-content/uploads/sites/9/media/Files/Reports/PIP_HealthOnline.pdf
- Fox, S. , & Fallows, D. (2003). Internet health resources Retrieved from https://www.pewinternet.org/wp-content/uploads/sites/9/media/Files/Reports/2003/PIP_Health_Report_July_2003.pdf.pdf
- Frenda, S. J. , Knowles, E. D. , Saletan, W. , & Loftus, E. F. (2013). False memories of fabricated political events. Journal of Experimental Social Psychology, 49(2), 280–286. 10.1016/j.jesp.2012.10.013 [DOI] [Google Scholar]
- Guess, A. , Nagler, J. , & Tucker, J. (2019). Less than you think: Prevalence and predictors of fake news dissemination on Facebook. Science Advances, 5(1), eaau4586. 10.1126/sciadv.aau4586 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hegeman, J . (2020). Providing people with additional context about content they share . Retrieved from https://about.fb.com/news/2020/06/more-context-for-news-articles-and-other-content/
- Lewandowsky, S. , Ecker, U. K. H. , Seifert, C. M. , Schwarz, N. , & Cook, J. (2012). Misinformation and its correction: Continued influence and successful debiasing. Psychological Science in the Public Interest, 13(3), 106–131. 10.1177/1529100612451018 [DOI] [PubMed] [Google Scholar]
- Loftus, E. F. (2005). Planting misinformation in the human mind: A 30‐year investigation of the malleability of memory. Learning & Memory, 12(4), 361–366. 10.1101/lm.94705 [DOI] [PubMed] [Google Scholar]
- Mitchell, A. , Gottfried, J. , Stocking, G. , Walker, M. , & Fedeli, S. (2019). Many Americans say made‐up news is a critical problem that needs to be fixed. Pew Research Center. Retrieved from https://www.journalism.org/2019/06/05/many-americans-say-made-up-news-is-a-critical-problem-that-needs-to-be-fixed/
- Patihis, L. , Frenda, S. J. , Leport, A. K. R. , Petersen, N. , Nichols, R. M. , Stark, C. E. L. , … Loftus, E. F. (2013). False memories in highly superior autobiographical memory individuals. Proceedings of the National Academy of Sciences of the United States of America, 20, 20947–20952. 10.1073/pnas.1314373110 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Pennycook, G. , Bear, A. , Collins, E. T. , & Rand, D. G. (2020). The implied truth effect: Attaching warnings to a subset of fake news headlines increases perceived accuracy of headlines without warnings. Management Science, (July). 10.1287/mnsc.2019.3478 [DOI]
- Pennycook, G. , McPhetres, J. , Zhang, Y. , Lu, J. G. , & Rand, D. G. (2020). Fighting COVID‐19 misinformation on social media: Experimental evidence for a scalable accuracy‐nudge intervention. Psychological Science, 31, 770–780. 10.1177/0956797620939054 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Roth, Y. , & Pickles, N. (2020). Updating our approach to misleading information. Retrieved from https://blog.twitter.com/en_us/topics/product/2020/updating-our-approach-to-misleading-information.html
- Smith, J. , Jackson, G. , & Raj, S. (2017). Designing against misinformation . Retrieved from https://medium.com/facebook-design/designing-against-misinformation-e5846b3aa1e2
- Southwell, B. G. , Niederdeppe, J. , Cappella, J. N. , Gaysynsky, A. , Kelley, D. E. , Oh, A. , … Chou, W.‐Y. S. (2019). Misinformation as a misunderstood challenge to public health. American Journal of Preventive Medicine, 57(2), 282–285. 10.1016/j.amepre.2019.03.009 [DOI] [PubMed] [Google Scholar]
- Sutherland, R. , & Hayne, H. (2001). Age‐related changes in the misinformation effect. Journal of Experimental Child Psychology, 79(4), 388–404. 10.1006/jecp.2000.2610 [DOI] [PubMed] [Google Scholar]
- Swire‐Thompson, B. , & Lazer, D. (2020). Public health and online misinformation: Challenges and recommendations. Annual Review of Public Health, 41(1), 433–451. 10.1146/annurev-publhealth-040119-094127 [DOI] [PubMed] [Google Scholar]
- United Nations Educational Scientific and Cultural Organization (2018). Journalism, “fake news” & disinformation. In Handbook for journalism education and training Retrieved from https://en.unesco.org/fightfakenews [Google Scholar]
- Walter, N. , & Tukachinsky, R. (2020). A meta‐analytic examination of the continued influence of misinformation in the face of correction: How powerful is it, why does it happen, and how to stop it? Communication Research, 47(2), 155–177. 10.1177/0093650219854600 [DOI] [Google Scholar]
- Wang, Y. , McKee, M. , Torbica, A. , & Stuckler, D. (2019). Systematic literature review on the spread of health‐related misinformation on social media. Social Science & Medicine, 240(September), 112552. 10.1016/j.socscimed.2019.112552 [DOI] [PMC free article] [PubMed] [Google Scholar]
- World Health Organization . (2020). Novel Coronavirus (2019‐nCoV): Situation report‐13 . Retrieved from https://www.who.int/docs/default-source/coronaviruse/situation-reports/20200202-sitrep-13-ncov-v3.pdf
- Wylie, L. E. , Patihis, L. , McCuller, L. L. , Davis, D. , Brank, E. , Loftus, E. F. , & Bornstein, B. (2014). Misinformation effect in older versus younger adults: A meta‐analysis and review. In Toglia M. P., Ross D. F., Pozzulo J., & Pica E. (Eds.), The elderly eyewitness in court. (pp. 38–66), London, England: Psychology Press. 10.4324/9781315813936 [DOI] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
No data were created or analyzed for the current essay.