Skip to main content
PLOS One logoLink to PLOS One
. 2022 Jul 18;17(7):e0271566. doi: 10.1371/journal.pone.0271566

Examining the role of information integration in the continued influence effect using an event segmentation approach

Jasmyne A Sanderson 1,*, Simon Farrell 1, Ullrich K H Ecker 1
Editor: Shane Lindsay2
PMCID: PMC9292086  PMID: 35849610

Abstract

Misinformation regarding the cause of an event often continues to influence an individual’s event-related reasoning, even after they have received a retraction. This is known as the continued influence effect (CIE). Dominant theoretical models of the CIE have suggested the effect arises primarily from failures to retrieve the correction. However, recent research has implicated information integration and memory updating processes in the CIE. As a behavioural test of integration, we applied an event segmentation approach to the CIE paradigm. Event segmentation theory suggests that incoming information is parsed into distinct events separated by event boundaries, which can have implications for memory. As such, when an individual encodes an event report that contains a retraction, the presence of event boundaries should impair retraction integration and memory updating, resulting in an enhanced CIE. Experiments 1 and 2 employed spatial event segmentation boundaries in an attempt to manipulate the ease with which a retraction can be integrated into a participant’s mental event model. While Experiment 1 showed no impact of an event boundary, Experiment 2 yielded evidence that an event boundary resulted in a reduced CIE. To the extent that this finding reflects enhanced retrieval of the retraction relative to the misinformation, it is more in line with retrieval accounts of the CIE.

Introduction

Misinformation regarding the cause of an event can continue to influence people’s event-related reasoning even if it has been retracted in a clear and credible manner by the issuing of a correction. This is known as the continued influence effect [CIE; 18]. The CIE has been demonstrated under controlled laboratory environments; for example, in a study by Johnson and Seifert [2], participants were given a news report detailing a fictitious warehouse fire. Participants were originally told that the fire was caused by negligent storage of volatile materials, but this was later retracted. Despite acknowledging the retraction, when answering questions about the fire, participants continued to rely on the initial incorrect information. Arguably, the CIE has also played a role in a number of real-world misinformation cases such as the MMR-autism scare, where despite strong evidence to the contrary, misinformation linking the MMR vaccine to autism had significant and lasting influence [3, 912].

Previous research has focused primarily on the role that memory retrieval plays for the CIE. Specifically, the CIE has been assumed to arise from selective retrieval of the misinformation, or failure to retrieve the correction. According to this view, memory representations compete for activation, and the CIE occurs if the misinformation is retrieved but its correction is not [1315]. A competing explanation stems from a mental-model account. This account proposes that people build mental models of events as they unfold [1618]. When new information relevant to the event is encountered, the mental model of the event needs to be updated. It has been argued that this updating process is particularly effortful if the new information features a correction that supersedes earlier information, because a correction requires the removal of old outdated information (i.e., the misinformation) and the encoding and integration of the new information to form a coherent updated model that reflects the current state of affairs [1922]. Therefore, if a correction is not sufficiently integrated during model updating, people may continue to rely on the outdated misinformation in their inferential reasoning.

Recent neuropsychological research has supported the mental-model account, indicating that the CIE may arise in part due to a failure in the integration and updating of information in memory [23]. Using functional magnetic resonance imaging, Gordon et al. [23] found differing neural activity for the encoding of retractions compared to non-retractions in regions involved in information integration (i.e., the precuneus and posterior cingulate gyrus), whereas no retraction-vs.-non-retraction difference was found during retrieval. This was taken to suggest that the CIE may arise due to a failure to integrate a retraction into a person’s mental model. However, a follow-up study failed to find neural differences at encoding; instead the results implicated a failure of retrieval processes [24].

It is therefore still unclear exactly what role integration and updating failure play in the CIE. This study expanded on previous research by investigating information integration and updating using a behavioural approach. Specifically, this study sought to manipulate the ease with which a retraction can be integrated with misinformation through use of event segmentation boundaries.

Event segmentation effects on integration and memory

As people go about their everyday life, they experience a continuous stream of incoming information, and yet this information is often remembered as distinct events or episodes in memory. There are a number of prominent models and theories of event cognition that can explain how this occurs [2527]. In particular, event segmentation theory [EST; 2830] and more recently the event horizon model [26] propose that individuals segment incoming information into discrete, meaningful events—a segment of activity that is perceived to have a defined beginning and end—which are separated by event boundaries. This segmentation process is thought to be automatic and triggered by contextual change (i.e., a change in one or more dimensions of the contextual environment such as spatial location). In other words, if context is perceived as constant, incoming information will continue to be interpreted as part of the same event; however, when there is a deviation in context, an event boundary is created, and thus two separate event representations [31]. This is consistent with the finding that individuals often identify event boundaries in text narratives and films at points where the features of a situation change [e.g., a character changes location; 29, 3235].

According to EST, events are represented in memory within an event model—a mental representation of the structure and content of the current event—which aids comprehension and is used to guide predictions about future behaviour and perceptual input. Event models are updated at event boundaries due to the increase in prediction error that occurs when aspects of the event change. EST argues for a global updating process, in which following an event boundary, the old event model is abandoned and a new model is created. The new model is then actively maintained in memory until the next boundary is encountered [36]. In this way, event segmentation acts as a control process that regulates and organises information in memory through updating at event boundaries [37]. This then has implications for the accessibility of information after a boundary has been crossed [34].

It follows that event segmentation can impact learning and memory [28, 30]. In particular, information from a specific event is better recalled while the event is still ongoing, as opposed to after the event, when an event boundary lies between the encoding of the information and its retrieval. Previous research has supported the effect of event boundaries on information accessibility using text narratives, film, and virtual-reality (VR) environments. These studies typically find that crossing an event boundary decreases the accessibility of information encoded prior to the boundary, as reflected in increased response times and reduced response accuracy, as well as slower reading times for boundary compared to non-boundary sentences [17, 28, 35, 3746]. These effects occur independently of pure temporal parameters such as the absolute amount of time since encoding or the temporal distinctiveness of the to-be-recalled information [47].

Spatial event segmentation

Changes in spatial context—which is generally an important dimension of memory representations [e.g., 26]—have been investigated as triggers of event segmentation [35, 48]. Shifts in spatial context (i.e., a change in location) may act as event-boundary markers that can subsequently disrupt cognitive processing and memory retrieval [26, 39, 49, see 50 for a review]. Consistent with this, Rinck and Bower [41] had participants memorise the layout of a building with specific objects located within each room, before reading a narrative about a character moving through the building. After the character moved to a new room, readers were slower at recognising objects from previous rooms. Similar results have been found for interactive environments.

The most prominent example of this is the location updating effect, which refers to the decline in memory that results when an agent moves from one location to another [e.g., passing through a doorway; e.g., 39, 51, 52]. In a series of experiments conducted by Radvansky and colleagues [39, 51, 52], participants navigated a series of rooms while picking up and setting down different objects in each room. During this, participants’ recognition memory for the objects was tested. The critical factor of interest was whether or not participants shifted to a new room between encoding and test. These studies found that memory for objects was less accurate and slower following a shift in location compared to when there was no shift. In other words, the accessibility of information from the previous room decreased after a spatial event boundary had been crossed. This finding was even more pronounced when participants moved through multiple rooms, therefore crossing multiple event boundaries [51]. The effect occurs in both VR [38, 39, 51, 52] and physical laboratory environments [51], as well as in people’s imagination [53]. It has been found when word pairs were used instead of objects [52], when the rooms were separated by transparent “glass” walls allowing participants to preview the next location [54], when recall was tested as opposed to recognition [55], and for both younger and older adults [56]. Furthermore, the effect does not appear to be simply due to context-dependent memory which would suggest that returning to the location where information was initially learned would improve recall [e.g., 57, 58], as memory was affected even when individuals were tested in the original room of encoding [51].

In summary, consistent with event segmentation theory, it appears that the presence of spatial event boundaries in either text narratives, film, or interactive environments influences the ease with which information is integrated and later accessed, with disruptions to memory occurring for information encoded prior to a boundary.

The current study

The current study investigated the effect of spatial event segmentation boundaries on information integration and updating in a CIE paradigm. According to the mental-model account and event segmentation theory, a spatial event boundary between the encoding of the misinformation and retraction should disrupt retraction integration and model updating. To illustrate, the work of Kendeou [21, 22, also see 5961] has shown that memory updating is facilitated when the initial misinformation and its subsequent correction are co-activated in memory and a conflict detected. This should be made less likely by an event boundary occurring between the two pieces of information, as encountering the correction may fail to trigger successful retrieval of the earlier misinformation. In turn, this could hinder integration and mental-model revision, thus resulting in increased misinformation reliance.

Across two experiments, misinformation and retractions were presented with or without a spatial event boundary between their encoding. To employ the spatial event boundary, participants either moved physically (Experiment 1) or virtually (Experiment 2) between spatial contexts. Physical/virtual spatial context changes were chosen so as to accommodate the event report format typically used in CIE studies, which do not lend themselves to narrative spatial shifts. For example, in contrast to the story-based narratives typically used in event segmentation studies, event reports are by nature less linear in terms of structure and less confined to a particular context/person (e.g., written in third person with input from multiple sources), which makes it difficult to implement event boundaries. Both experiments included inference questions to assess misinformation reliance.

Experiment 1

The aim of Experiment 1 was to investigate the role of information integration and updating in the CIE. Specifically, it aimed to examine the effect of a physical spatial event boundary on the integration of misinformation and its retraction, applying an event segmentation approach to the CIE paradigm. We manipulated the presence versus absence of a spatial event boundary, via a shift in spatial context, between misinformation and retraction encoding. Participants either read both the misinformation and retraction in the same spatial context (i.e., the same room; no-shift condition) or in two different contexts by shifting locations to a new room before reading the retraction (shift condition). Participants then answered a series of inferential reasoning questions (in a third room) to determine the effect of the boundary on retraction integration and subsequent misinformation reliance.

There were two hypotheses: The main hypothesis was that integration would be facilitated in the no-shift condition—when misinformation and retraction were encoded in the same spatial context—and that this would reduce subsequent misinformation reliance, relative to the shift condition in which there was a spatial event boundary between misinformation and retraction encoding. A secondary hypothesis was that a retraction (in either shift or no-shift conditions) would reduce reliance on misinformation compared to a no-retraction control, without entirely eliminating its influence, replicating previous work [e.g., 15, 62, 63].

Method

Experiment 1 used a between-subjects design contrasting no-shift and shift conditions. Due to the resource-intensive nature of testing, it was decided to just implement these two conditions, while obtaining data from a no-retraction control condition from a separate sample of online participants, simply to obtain a rough baseline for comparison as per the secondary hypothesis. The dependent variable was participants’ inference score derived from their responses to post-manipulation inference questions.

Participants

The original target sample size was derived using a-priori power analysis [using G*Power 3; 64, also see 65] which suggested that to detect a difference between no-shift and shift conditions inference scores in a t-test, an effect of f = .20 (α = .05; 1-β = .80) would require a sample of 200 participants. The effect size of f = .20 was chosen based on Brysbaert [65] who recommends f = .20 as a default effect size for power analyses in psychological research. In addition, previous research [41] found an effect of f = .23 when comparing no-boundary and boundary conditions using an ANOVA. Due to difficulties in recruiting participants, it was recognised that a sequential Bayesian approach would be appropriate to balance the resource-intensive nature of testing against the evidence gained. Accordingly, the data were subjected to Bayesian sequential testing after testing half the number of planned participants, and committing to Bayesian statistics for inference beyond that point [66, 67]. As described below, preliminary results provided clear evidence for the null, and the experiment was stopped so as to not waste resources.

A total of N = 112 first-year undergraduate students from the University of Western Australia participated in exchange for course credit. The total sample comprised 77 females and 35 males; mean age was M = 20.83 years (SD = 5.87), ranging from 17 to 49 years. Participants were randomly assigned to either the no-shift or shift condition (with the constraint of roughly equal participant numbers; n = 56 in no-shift condition; n = 56 in shift condition).

A separate sample of n = 56 participants were recruited on Amazon MTurk [68] for the no-retraction control condition. Participants were U.S. residents; the sample comprised 26 females and 30 males; mean age was M = 37.23 years (SD = 10.61), ranging from 21 to 64 years. Participants were reimbursed US$1.80 through MTurk.

Materials

Event reports

Four event reports detailing fictitious scenarios (e.g., an emergency airplane landing) were used in the present study (although participants received only one out of the pool of four). These reports were based on those from Ecker et al. [62]. Each report comprised 12 individual messages and was separated into two parts—Part 1 (messages 1–6) contained the critical information (e.g., that the emergency landing was due to extreme weather conditions), and Part 2 (messages 7–12) contained either the retraction (e.g., that initial reports were incorrect and the landing was due to mechanical failure) or a neutral filler message (e.g., that the airline released a statement) in the no-retraction control condition. It was decided to not explicitly refer to the misinformation within the retraction as previous research [62] has found that explicit retractions were more effective in reducing misinformation reliance, presumably by facilitating conflict detection and model updating. As such, referring explicitly back to the misinformation may facilitate integration, which may reduce the impact of the event boundary manipulation. See S1 File for the full event reports.

Distractor tasks

Two distractor tasks were used. The first was a word finding task, which was completed between the two parts of the event report. This was used to achieve comparable time intervals between misinformation and retraction encoding across conditions. To this end, the distractor task duration was 2 min in the no-shift condition (and the control condition) but only 1 min in the shift-condition (to allow time for the move from one room to the other). The second distractor task was a word fluency task completed after the event report. Six letters were presented successively on the screen for 30 s each (i.e., 3 min total), and participants had to type as many words as possible starting with each letter.

Questionnaires

There was one questionnaire specific to each report. Each questionnaire consisted of 11 questions that related to the relevant event report: an open-ended inference question, four memory questions and six rating scale inference questions. The questionnaires are provided in the S1 File.

The four memory questions assessed memory for details mentioned in the report to determine that participants were paying attention during encoding. These questions were presented in multiple-choice format with four response options (e.g., What airport did the airplane land at?–a. Portland; b. Denver; c. Orlando; d. Seattle). Each question assessed a different part of the report (i.e., before and after the critical information, before and after the retraction/filler statement).

Seven inference questions were used to assess participants’ reliance on the critical information. The first question was an open-ended question that asked participants to briefly summarise the report. The remaining six were statements relating to the critical information (e.g., The US guidelines for flying in bad weather should be reviewed). Agreement with each statement was rated on an 11-point Likert scale ranging from “Completely disagree” (0) to “Completely agree” (10). One of these statements was negatively coded.

There was an additional question at the end of the experiment asking participants whether they had put in a reasonable effort (participants answering “no” were excluded from analysis; see below for details).

Rooms

The event reports were read in one or two rooms, depending on condition. The rooms and respective survey versions (see below) were set up to be as different as possible (see Fig 1). Room A was decorated with four landscape posters on adjacent walls, dim lighting via a lamp, an air freshener, items on the desk (e.g., plant, scent diffusor, tissues), and an office chair with wheels. The Room A survey was presented on a small square monitor (16-inch LG Flatron L1919S); it used a black background and the text was white Georgia size 12 font. Room B had plain white walls, overhead lighting, no items on the desk, and a padded steel frame chair without wheels. The Room B survey was presented on a large rectangular monitor (21.5-inch ViewSonic VX2252mh); it used a white background with blue Arial size 16 font. Room C was a communal lab space with four workstations set up, each with a computer monitor and office chair with wheels; one wall contained a bookshelf and another a window.

Fig 1. Rooms used in Experiment 1.

Fig 1

Left: Room A; middle: Room B; right: Room C.

Procedure

Ethics approval was granted by the University of Western Australia’s Human Research Ethics Office. Participants received an approved information sheet and provided informed consent prior to participating for this and all other experiments. Participants read one event report (out of the pool of four). The report was presented one message at a time on a computer screen, delivered via Qualtrics software (Qualtrics, Provo, UT, USA). The rooms and event report topics were counterbalanced across participants.

In the no-shift condition, participants read both Part 1 and Part 2 of the event report in the same room (Room A or B). The word finding distractor task was completed for 2 minutes in between the two parts of the report in the same room. In the shift-condition, participants read the first part of the event report in one room (Room A or B) and then moved to a second room (Room B or A respectively) to complete the word finding distractor task for 1 minute and read the second part of the report. All participants, regardless of condition, then moved to Room C, where they completed the word fluency distractor task for 3 minutes and answered the relevant pen-and-paper questionnaire (see Fig 2). Participants were then fully debriefed. The experiment took approximately 15 minutes. Apart from the fact that participants completed the experiment online (and thus did not move locations during the survey), the procedure in the control condition was identical.

Fig 2. Schematic depiction of Experiment 1’s no-shift and shift conditions procedure.

Fig 2

Note: Pt 1, part one of the event report; Pt 2, part two of the event report.

Scoring

Memory scores

Memory scores were calculated from responses to the four multiple-choice memory questions. Correct responses were given a score of 1 and incorrect responses received a score of 0, resulting in a possible maximum memory score of four.

Inference scores

Inference scores were calculated from responses to the open-ended question and the six rating-scale items. The negatively worded items were reverse-scored. Responses to the open-ended question were coded by the experimenter for recall of the critical information, the retraction, and the provided alternative explanation. A score of 1 was given when the critical information, retraction, or alternative was recalled and 0 otherwise. Thus, participants were able to receive a score of 1 on all three scores (e.g., recall of the critical information: “it was stated the emergency landing was due to bad weather”, the retraction: “as it turned out, it was not actually due to the weather”, and the alternative: “it was due to a mechanical fault”). The no-retraction control condition was not coded for recall of the retraction or alternative. The coded scores were then converted into an inference score of 0 or 10 to maintain the same scaling as the rating scale inference questions: If participants recalled the retraction and/or the alternative, they received a score of 0 (indicating no misinformation reliance) irrespective of whether they also recalled the critical information; if participants recalled only the critical information without the retraction or alternative, they received a score of 10 (indicating misinformation reliance). A total inference score was then calculated as the average of the seven inference questions; higher inference scores indicate greater misinformation reliance.

Results

One participant in the control condition failed all memory questions; their data were excluded. Additionally, one participant in the no-shift condition was excluded who indicated that their data should not be used due to lack of effort. Finally, one participant in the shift condition was excluded due to failure to complete the experiment. Thus, the final sample was N = 165 (n = 55 per condition).

All Bayesian analyses were conducted in JASP, using default priors (fixed-effects scale parameter r = 0.5; random-effects scale parameter r = 1). Bayes factors (BF) represent the relative evidence for one model over another, given the data. Any BF can be expressed as either BF10, which quantifies support for the alternative hypothesis, or BF01, which quantifies support for the null hypothesis. A BF in the range 1–3 (1/3-1) provides anecdotal evidence, a BF of 3–10 (1/10-1/3) provides substantial evidence, a BF of 10–30 (1/30-1/10) provides strong evidence, a BF of 30–100 (1/100-1/30) provides very strong evidence, and a BF > 100 (< 1/100) constitutes extreme evidence [69].

Memory scores

The mean memory scores across the four event scenarios and the three conditions are shown in Table 1.

Table 1. Descriptive statistics for memory scores across scenarios and conditions.

Airplane Landing Bushfire Water Source Nightclub
M SE M SE M SE M SE
Control 2.80 .20 3.53 .19 3.23 .23 3.50 .20
No-shift 3.43 .17 3.62 .14 2.92 .24 3.73 .12
Shift 3.64 .13 3.29 .22 3.00 .20 3.86 .14

The mean memory scores were comparable across control (C), no-shift (NS), and shift (S) conditions; mean scores were MC = 3.26 (SEC = .11), MNS = 3.44 (SENS = .09), and MS = 3.46 (SES = .10). A Bayesian independent samples t-test provided substantial evidence against a difference between no-shift and shift conditions, t(108) = -.14, BF01 = 4.91.

Inference scores

Mean inference scores were MC = 6.50 (SEC = .27), MNS = 4.00 (SENS = .32), and MS = 3.57 (SES = .25). The mean inference scores across the four event scenarios and the three conditions are shown in Table 2.

Table 2. Descriptive statistics for inference scores across scenarios and conditions.
Min Max M SE
Airplane Landing
    Control 0.86 9.57 5.89 0.53
    No-shift 3.29 7.57 5.70 0.26
    Shift 2.00 7.43 4.58 0.45
Bushfire
    Control 3.29 8.71 5.43 0.41
    No-shift 0.00 5.86 2.26 0.58
    Shift 1.43 5.57 3.69 0.35
Water Source
    Control 3.29 10.00 7.07 0.55
    No-shift 1.86 8.29 5.01 0.56
    Shift 1.43 7.29 3.70 0.49
Nightclub
    Control 5.43 9.86 7.99 0.37
    No-shift 0.00 8.14 3.03 0.65
    Shift 0.00 6.86 2.32 0.51

Numerically, both no-shift and shift conditions had lower inference scores than the no-retraction control, suggesting some effect of the retraction. Two Bayesian one-sample t-tests determined that inference scores in both no-shift and shift conditions were substantially different from zero, ts(54) ≥ 12.36, BF10 ≥ 1.68e+17, indicating the presence of a CIE. We acknowledge that zero may not be an appropriate baseline when using rating scales, but use zero in the absence of any other obvious baseline. A Bayesian independent samples t-test (with default priors: fixed-effects scale parameter r = 0.5; random-effects scale parameter r = 1) comparing no-shift and shift conditions yielded substantial evidence against a main effect of condition on participants’ reliance misinformation in their inferential reasoning, t(108) = 1.05, BF01 = 3.02 (though numerically the no-shift condition had a higher inference score than the shift condition). In addition, a one-tailed test (specifying H1 as NS < S) was run to quantify the evidence against the directional hypothesis. This provided strong evidence against H1, t(108) = 1.05, BF01 = 9.39.

Discussion

To investigate the potential role of information integration in the CIE, Experiment 1 examined whether the presence of a spatial event boundary would impair the integration of a retraction. It was predicted that integration would be facilitated when misinformation and retraction were encoded in the same spatial context, and that this would reduce misinformation reliance relative to a condition with a spatial event boundary between misinformation and retraction encoding. The observed data indicated that a retraction in both no-shift and shift conditions reduced reliance on misinformation (without entirely eliminating it) compared to a no-retraction control condition, consistent with prior research [15, 62, 63]. However, contrary to predictions, a spatial event boundary did not have a substantial effect on the integration of a retraction, at least not to the extent that the effect was measurable in participants’ misinformation reliance. While numerically there was an effect, it was in the direction opposite of that predicted, with greater reliance on misinformation in the no-shift condition.

This finding is inconsistent with the EST assumption that event boundaries induced through spatial context shifts disrupt access to information from the preceding event. Possible explanations for why Experiment 1 failed to find evidence for an effect of a spatial event boundary relate to (1) the differences in cognitive processing relative to previous EST studies, (2) the causal structure of the event reports used, and (3) the spatial context manipulation itself.

Firstly, Experiment 1 did not measure retraction integration or retrieval directly. Instead, it used a CIE paradigm, which measures different cognitive processes than the typical paradigms used in EST research. As far as we are aware, this is the first study to investigate the impact of event boundaries in a CIE paradigm. In previous EST research, the effect of event segmentation boundaries has been demonstrated on memory retrieval directly. Based on this prior research, we assumed that segmentation boundaries would reduce mnemonic accessibility of the misinformation, which would in turn impair information integration, resulting in stronger continued influence of the misinformation on inferential reasoning. Therefore, there were two additional processing steps (i.e., integration and continued influence) that had not been previously considered by EST studies. As such, one explanation for the inconsistent results may be that the boundary affected mnemonic accessibility in line with previous research, without this impacting on information integration and/or inferential reasoning downstream. It is also possible that the reasoning measures typically applied in CIE studies are not sensitive enough to detect potential boundary effects.

Secondly, the reports typically used in the EST literature are narratives with a linear timeline (e.g., Jan did this, she then did this, she then did that…). The event reports used in Experiment 1 as well as previous CIE research are by nature less linear (e.g., the events described have time gaps or reversals) and have an underlying causal structure (e.g., X caused Y…actually it was Z that caused Y), which may have negated the effect of the boundary. The event horizon model suggests that individuals monitor causal information in events, and that causal connections can act as a way of linking events, which may thus influence whether an event is interpreted as being a part of the same or different event model [7072]. That is, causally related information is more likely than unrelated information to be interpreted as belonging to the same event, irrespective of the underlying temporal structure of the report [35, 70, 71]. Therefore, the causal structure of the event reports may have influenced whether the spatial boundary manipulation had an effect on the integration of the retraction. In addition, it might be speculated that the inclusion of response alternatives in the memory questions may have facilitated general recall of the event, thereby indirectly reducing the impact of the boundary manipulation on inferential reasoning.

Thirdly, it may be that the rooms were not as contextually different as was required to find spatial event segmentation effects. While every attempt was made to make the rooms different in a perceptually salient manner, there were practicality constraints, as the rooms used for encoding were both testing booths with similar layouts (e.g., small rooms containing a desk against one wall) within a larger lab space. It may have been that rather than interpreting the two rooms as two different spatial locations, participants may have interpreted them as being elements of one larger location (i.e., the lab), resulting in only one event model. Stronger manipulations, for example through use of a VR environment, may allow a greater distinction between the spatial locations. Alternatively, it may be the case that the rooms were sufficiently different but that the spatial context manipulation lacked task relevance. Some studies have failed to produce spatial boundary effects with narrative texts unless participants were specifically instructed to monitor spatial context [7377]. As such, it has been suggested that if spatial context information is not in an individual’s attentional focus (i.e., emphasised by or important to the text or task), it may be less likely to be noticed and encoded [39, 71, 77, 78]. This suggests that people may not necessarily construct event models based on spatial location unless it is made salient or functionally important to the task [74]. In addition, the spatial context change occurred outside the narrative environment (as opposed to within the narrative; e.g., a character changing locations) and it may be that in order for participants to perceive the boundary as relevant to the narrative there needed to be both narrative and extra-narrative spatial shifts [see 79, for a similar argument].

Experiment 2

Experiment 2 sought to extend Experiment 1 by using a VR environment to better differentiate between the locations, and to make the spatial shift more central to the memoranda. Participants played the role of a detective who read clues with the expectation that the clues would be needed for a later task in the story setting. The clues related to a fictional crime, and bore on the likely culpability of a main suspect. The misinformation implicated the main suspect, with other clues not bearing directly on the guilt of that suspect. Participants either read both the misinformation and retraction in the same spatial context (i.e., a living room depicting a crime scene; no-shift condition) or in two different contexts by shifting locations to a police station foyer before reading the retraction (shift condition). Participants then answered a series of inferential reasoning questions (in a third room; a police interrogation room) to determine the effect of the boundary on retraction integration and subsequent misinformation reliance.

Method

Experiment 2 used a between-subjects design contrasting no-shift and shift conditions. Due to the resource-intensive nature of testing, it was decided to implement these two core conditions only. The dependent variables were participants’ inference scores derived from their responses to the post-manipulation inference questions, and a “charge” measure based on a question asking participants which suspect they would charge. The design and analysis plan were pre-registered (https://osf.io/kg73n).

Participants

We anticipated that a sample size of 100 would be feasible, given practicality constraints; this was the pre-registered sample size. A total of 127 students from the University of Western Australia and community members participated. No demographics were recorded. Participants were randomly assigned to either the no-shift or shift condition (with the constraint of roughly equal cell numbers; n = 64 in no-shift condition; n = 63 in shift condition).

Apparatus

VR equipment

We used a HTC Vive Pro headset, a controller, and two tripod base stations. The task was programmed in Unity and run via SteamVR on a Gigabyte Sabre 15 Laptop, with a resolution of 1920 x 1080 pixels.

Materials

Event report

The event report was displayed as a series of messages that became available at certain points during the participants’ exploration of the VR environment (see details below). The report detailed a house robbery, and was based loosely on Johnson and Seifert’s [2] jewellery theft narrative. The report was separated into two parts. Part 1 contained a general explanatory paragraph setting the scene (i.e., the participant was told they are a detective who has arrived at a crime scene to investigate a burglary), followed by eight clues about the crime (e.g., a blue thread potentially originating from the resident’s scarf), which included the critical information (i.e., a clue that implicated a particular suspect, viz. the homeowner’s son Evan; e.g., a bookmaker’s betting slip and reports from the family indicate Evan has a gambling problem and has accumulated large debts). Part 2 of the report contained eight retraction statements presented as emails, which retracted both non-critical information (e.g., clarifying that the thread did not belong to the homeowner’s scarf) and the critical information (i.e., several sources confirmed that the son was out of town). Retractions of non-critical information were included in order to not overly highlight the retraction of the critical information, and to maintain consistency with previous CIE research, which tends to present retractions not in isolation but in the context of other filler information [e.g., 80, 81]. See S1 File for the full event report.

Questionnaire

The questionnaire consisted of eleven questions that related to the event report: four inference questions, one charge question, and six memory questions. All questions are provided in the S1 File.

The six memory questions assessed memory for details mentioned in the report, to determine that participants were paying attention during encoding. These questions were presented in multiple-choice format with four response options (e.g., What colour was the button?—a. Blue; b. Green; c. Red; d. Yellow).

The four inference questions were used to assess participants’ reliance on the critical information. These questions were presented as statements relating to the critical information (e.g., How likely are the family to be angry with their son?). Agreement with each statement was rated on an 11-point Likert scale ranging from 0 to 10.

The charge question was a multiple-choice question asking participants who they were going to charge (i.e., Who will you charge?—a. Evan [son]; b. H. Brown [known offender]; c. Nobody [insufficient evidence]). This was included as an exploratory measure only.

VR rooms

Participants stood in a virtual environment, which was one of three locations: the crime scene, the police station foyer, and the interrogation room (see Fig 3). The crime scene location was designed as a living room (i.e., warm lighting, lounge, sliding doors), which was markedly different from the police station foyer (i.e., bright lights, officer sitting behind a glass divider). The interrogation room was a dimly lit room and contained a metal desk. In the VR environment, a controller was visible (its position and orientation linked to that of the physical controller held by the participant) and emitted a laser to show where it was pointing. The bottom trigger of the controller was used to record a response.

Fig 3. Rooms used in Experiment 2.

Fig 3

Left: Crime Scene; middle: Foyer; right: Interrogation Room.

Procedure

Participants were tested individually in the lab. They first received general task instructions (including cautions about use of VR equipment) and a cover story, before putting on the VR set. Participants remained standing for the entirety of the task with their back to a wall. The experimenter adjusted the VR headset and the inter-display distance until the participant reported that the environment was in focus (not blurry) and comfortable. Participants first completed a familiarisation phase in the living room, where they practiced looking around and using the controller in the VR environment and responding to prompts.

Participants then received a more detailed introduction to the crime scene and explored the crime scene. Participants were able to inspect clues by finding and clicking on a magnifying glass. The magnifying glass for each clue appeared at 10 s intervals; participants were able to find clues before the magnifying glass appeared; however, participants rarely managed to do this given the timing. Participant could inspect the clues for as long as they liked. Clues were queued such that a clue would not appear until the previous clue had disappeared. As a consequence, the clues in the living room in Part 1 were presented in a fixed order.

Once all clues had been provided, the information from Part 2 was presented, which involved retractions of both non-critical and critical information. Each item in Part 2 was presented as an email in the VR environment; an example is shown in Fig 4. There was a 15 s interval between the last clue and the first email. Participants read each email and then moved on to the next email by clicking the ‘Next’ button.

Fig 4. Example of Experiment 2’s retraction email in the shift condition.

Fig 4

In the no-shift condition, participants read both Part 1 and Part 2 of the event report at the crime scene and then moved to the police station for a brief period, where they were asked to wait briefly until the interrogation room was available. In the shift-condition, participants read Part 1 at the crime scene and then moved to the police station to read Part 2. Following a 10 s interval, all participants then moved to the interrogation room, where they received the questionnaire. The timing was more or less controlled between conditions (although participants determined the timing of clues and emails by how fast they read and responded).

On completion of the experiment, participants were fully debriefed. The experiment took approximately 15 minutes.

Scoring

Memory scores

Memory scores were calculated from responses to the six multiple-choice memory questions. Correct responses were given a score of 1 and incorrect responses received a score of 0, resulting in a possible maximum memory score of six.

Inference scores

Inference scores were calculated by averaging responses to the four rating-scale items. The negatively worded item was reverse-scored. Higher inference scores indicate greater misinformation reliance, with a maximum possible score of 10.

Charge scores

Charge scores were based on the multiple-choice charge question. Responses were converted into a score of 0 or 10 (to maintain consistent scaling with the rating scale inference items): If participants chose Evan (the son), they received a score of 10 (indicating misinformation reliance); if participants chose nobody or an alternative suspect, they received a score of 0 (indicating no misinformation reliance).

Results

Based on pre-registered a-priori minimum-performance outlier criteria, two participants with poor memory for the materials (defined as less than 50% correct) were excluded from analysis, yielding a final sample of N = 125 (n = 64 in no-shift condition; n = 61 in shift condition). In line with our pre-registration, analyses were conducted on N = 100 (the first 100 completions; n = 52 in the no-shift condition; n = 48 in the shift condition). However, additional analyses conducted on the full sample (N = 125) were comparable and are provided.

Memory scores

The mean memory scores were comparable across no-shift (MNS = 5.12; SENS = .13) and shift (MS = 5.06, SES = .14) conditions. A Bayesian independent-samples t-test provided substantial evidence against a difference between conditions in participants’ memory for the report, t(98) = .28, BF01 = 4.58.

With N = 125, mean memory scores were MNS = 5.08, SENS = .12 and MS = 4.98, SES = .13. A Bayesian independent-samples t-test provided substantial evidence against a difference between conditions, t(123) = .54, BF01 = 4.59.

Inference scores

Mean inference scores were MNS = 4.24 (SENS = .22), and MS = 3.43 (SES = .29). Two Bayesian one-sample t-tests determined that no-shift and shift condition inference scores were substantially different from zero, ts(47/51) ≥ 11.91, BF10 ≥ 3.92e+21, indicating the presence of a CIE. To test the hypothesis that misinformation reliance would be lower in the no-shift condition compared to the shift condition, an undirected (two-tailed) Bayesian independent-samples t-test was run on inference scores. This provided anecdotal evidence for a difference between conditions, t(98) = 2.25, BF10 = 1.95, but in the direction opposite to predictions. In addition, a one-tailed test (specifying H1 as NS < S) was run to quantify the evidence against the directional hypothesis. This provided strong evidence against H1, t(98) = 2.25, BF01 = 14.36, reflecting the finding that contrary to our hypothesis, inference scores were greater in the no-shift compared to the shift condition.

With N = 125, mean inference scores were MNS = 4.29 (SENS = .20) and MS = 3.63 (SES = .24). Bayesian one-sample t-tests revealed decisive evidence that both conditions were greater than zero, ts(60/63) ≥ 14.85, BF10 ≥ 1.759e+27. A Bayesian independent-samples t-test found inference scores were greater in the no-shift compared to the shift condition, t(123) = 2.11, BF01 = 15.36 (one-tailed).

Charge scores

The mean charge score across the no-shift and shift conditions was MNS = 2.12 (SENS = .57), and MS = 1.88 (SES = .57). Two Bayesian chi-square tests were conducted to determine if the conditions charge scores differed from zero (zero indicating no misinformation reliance). Both conditions’ charge scores were substantially greater than zero, all χ2 (1) > 17.31, BF01 < 7.109e-4; indicating the presence of a CIE.

To test the hypothesis that the no-shift condition would have lower misinformation reliance than the shift condition, a Bayesian chi-square test (using the default prior: a = 1) with condition as the sole factor was conducted on participants’ charge scores. There was substantial evidence against a main effect of condition on participants’ decision to charge a suspect, χ2 (1) = .09, BF01 = 4.86.

With N = 125, mean charge scores were MNS = 2.03 (SENS = .51), and MS = 1.97 (SES = .51). Two Bayesian chi-square tests revealed substantial evidence that both conditions were greater than zero, all χ2 (1) > 22.44, BF01 < 4.684e-5. A Bayesian chi-square test suggested substantial evidence of no difference between conditions, χ2(1) = .01, BF01 = 5.63.

Discussion

The aim of Experiment 2 was to investigate the effect of a spatial event boundary on the integration of a retraction, and subsequent reliance on misinformation, using a VR environment. From the information integration approach it was predicted that integration would be facilitated in the no-shift condition and that this would reduce misinformation reliance, relative to the shift condition [22, 23, 39, 51, 52]. There appeared to be an effect of a spatial event boundary on participants’ inferential reasoning; however, the direction of this effect was opposite to our prediction, with greater misinformation reliance occurring in the no-shift condition compared to the shift condition.

At first glance, this result appears inconsistent with previous event segmentation literature, specifically research on the location updating effect, which has reported that shifting locations (i.e., crossing a spatial event boundary) negatively impacts memory compared to a no-shift condition [e.g., 51, 82]. However, a study by Pettijohn, Thompson, Tamplin, Krawietz, and Radvansky [83] found that event boundaries can occasionally improve memory when they add structure and organisation to the contents in memory, such that the segregation of information into different event models may reduce interference from competing information. Therefore, the boundary between misinformation and retraction may have helped participants in the shift condition retrieve the retraction without interference from the misinformation. In other words, the boundary may have impaired memory for the misinformation relative to memory for the retraction, which in turn reduced accessibility of and reliance on misinformation and thus improved inferential reasoning. The current study’s findings are broadly consistent with Experiment 1, inasmuch as Experiment 1 numerically also found greater reliance on misinformation in the no-shift condition and neither experiment yielded evidence for better integration in the no-shift condition.

We also note that due to the no-shift/shift manipulation, the number of spatial boundaries between the retraction and the test varied; specifically, the shift condition had one boundary between retraction and test (misinformation-shift-retraction-shift-test), whereas the no-shift condition had two (misinformation-retraction-shift-shift-test). Therefore, it may be that in the no-shift condition, memory for the retraction was impaired because there were two context shifts between retraction encoding and retrieval at test, compared to only one in the shift condition. From a temporal context view, this would mean that the relative contextual recency of the retraction (vs. misinformation) in the shift condition would reduce reliance on the misinformation [47, 84]. Indeed, previous misinformation research has noted the importance of recency when providing corrections based on the assumption that recent information is more easily retrieved [8587, also see 88, 89].

General discussion

The present research aimed to investigate the role of information integration in the continued influence effect (CIE). Specifically, Experiments 1 and 2 examined the effect of a spatial event boundary on the integration of a retraction, using physical and virtual reality environments.

The results of Experiment 1 suggested that a retraction was effective in reducing reliance on misinformation, as reflected in participants’ inferential reasoning scores, consistent with previous research [1, 5, 15, 62, 63]. However, while retractions appeared effective, both studies found that they did not completely eliminate reliance on misinformation, providing evidence for a CIE, again in line with previous research [2, 6, 14, 20, 90]. Future research could include a no-misinformation condition in order to formally establish the presence of a CIE.

Across both experiments, we failed to find evidence of a detrimental effect of a spatial event boundary on retraction integration, as reflected in similar inferential reasoning scores between no-shift and shift conditions in Experiment 1 and greater inference scores in the no-shift condition in Experiment 2. We acknowledge that while we can rule out a substantial effect, a small effect may be compatible with the observed results. However, the theoretical and practical relevance of such a small effect are questionable. This is inconsistent with the mental-model account of the CIE, which suggests that people may continue to rely on outdated misinformation in their inferential reasoning if a retraction is not sufficiently integrated during model updating [22, 23]. It was hypothesized that integration of a retraction should be facilitated in the no-shift condition (where the misinformation and its retraction should be part of the same event segment) compared to the shift condition (where the misinformation and its retraction should be part of different event segments), which in turn should result in lower, not greater, post-retraction misinformation reliance. Thus, overall, results provide some evidence against the mental-model account and the role of integration failure in the continued influence effect.

Overall, the results are more in line with the retrieval account of the CIE. This holds that both misinformation and retraction information are stored separately and potentially retrieved at the point of inference, with the reliance on misinformation depending on the ability to retrieve retraction information. This theory provides an explanation for the surprising finding of the lower misinformation reliance in the shift condition in Experiment 2, where the fewer spatial context shifts between the retraction and test may have facilitated retrieval of the retraction relative to the misinformation [also see 24, 81, 91].

Together, these studies have implications for event segmentation theory and the replicability and generalizability of event segmentation effects to a CIE context. In particular, the findings show that event segmentation may only impact specific retrieval processes but not downstream processes such as integration and updating. Future research may wish to investigate the effect of spatial event boundaries within the event reports (e.g., a protagonist moving locations), as opposed to being applied externally to the participant [41, 77]. Future research could also use other types of event boundaries (e.g., shifts in temporal context) [92].

Supporting information

S1 File. Supplemental information for Experiments 1 and 2.

(DOCX)

Acknowledgments

We thank Ryan Li for Experiment 2’s programming, and Charles Hanich for research assistance for Experiment 1 and 2. We also acknowledge the involvement of the PSYC3310 Topic 7 class of 2019 in some testing for Experiment 2.

Data Availability

Materials are provided in the Supplement. Data are available on the Open Science Framework website at https://osf.io/qrpjb/ (doi:10.17605/OSF.IO/QRPJB; Experiment 1) and https://osf.io/sxh2w/ (doi:10.17605/OSF.IO/SXH2W; Experiment 2).

Funding Statement

This research was supported by an Australian Government Research Training Program Scholarship and a Bruce and Betty Green Postgraduate Research Scholarship to JS. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Chan MPS, Jones CR, Hall Jamieson K, Albarracín D. Debunking: A meta-analysis of the psychological efficacy of messages countering misinformation. Psychol Sci. 2017;28:1531–46. doi: 10.1177/0956797617714579 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Johnson HM, Seifert CM. Sources of the continued influence effect: When misinformation in memory affects later inferences. J Exp Psychol Learn Mem Cogn. 1994;20(6):1420–36. 10.1037/0278-7393.20.6.1420 [DOI] [Google Scholar]
  • 3.Lewandowsky S, Ecker UKH, Seifert CM, Schwarz N, Cook J. Misinformation and its correction: Continued influence and successful debiasing. Psychol Sci Public Interest. 2012;13(3):106–31. doi: 10.1177/1529100612451018 [DOI] [PubMed] [Google Scholar]
  • 4.Sanderson JA, Ecker UKH. The Challenge of Misinformation and Ways to Reduce Its Impact. In: Van Meter P, List A, Lombardi D, Kendeou P, editors. The Handbook of Learning from Multiple Representations and Multiple Perspectives. New York, NY: Routledge; 2020. p. 461–76. [Google Scholar]
  • 5.Walter N, Murphy ST. How to unring the bell: A meta-analytic approach to correction of misinformation. Commun Monogr. 2018:1–19. 10.1080/03637751.2018.1467564 [DOI] [Google Scholar]
  • 6.Walter N, Tukachinsky R. A meta-analytic examination of the continued influence of misinformation in the face of correction: how powerful is it, why does it happen, and how to stop it? Commun Res. 2020;47:155–77. 10.1177/0093650219854600 [DOI] [Google Scholar]
  • 7.Ecker UK, Lewandowsky S, Cook J, Schmid P, Fazio LK, Brashier NM, et al. The psychological drivers of misinformation belief and its resistance to correction. Nat Rev Psy. 2022;1(1):13–29. [Google Scholar]
  • 8.Kan IP, Pizzonia KL, Drummey AB, Mikkelsen EJV. Exploring factors that mitigate the continued influence of misinformation. CR:PI. 2021;6(76):1–33. doi: 10.1186/s41235-021-00335-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Demicheli V, Jefferson T, Rivetti A, Price D. Vaccines for measles, mumps and rubella in children. Cochrane Database Syst Rev. 2005;2:1–161. doi: 10.1002/14651858.CD004407.pub2 [DOI] [PubMed] [Google Scholar]
  • 10.Smith MJ, Ellenberg SS, Bell LM, Rubin DM. Media coverage of the measles-mumps-rubella vaccine and autism controversy and its relationship to MMR immunization rates in the United States. Pediatrics. 2008;121(836–843). doi: 10.1542/peds.2007-1760 [DOI] [PubMed] [Google Scholar]
  • 11.Lewandowsky S, Stritzke WG, Freund AM, Oberauer K, Krueger JI. Misinformation, disinformation, and violent conflict: From Iraq and the “War on Terror” to future threats to peace. Am Psychol. 2013;68:487–501. doi: 10.1037/a0034515 [DOI] [PubMed] [Google Scholar]
  • 12.Nyhan B, Reifler J, Richey S, Freed GL. Effective messages in vaccine promotion: a randomized trial. Pediatrics. 2014;133:1–8. doi: 10.1542/peds.2013-2365 [DOI] [PubMed] [Google Scholar]
  • 13.Ayers MS, Reder LM. A theoretical review of the misinformation effect: Predictions from an activation-based memory model. Psychon Bull Rev. 1998;5(1):1–21. 10.3758/BF03209454 [DOI] [Google Scholar]
  • 14.Ecker UKH, Lewandowsky S, Tang DTW. Explicit warnings reduce but do not eliminate the continued influence of misinformation. Mem Cognit. 2010;38(8):1087–100. doi: 10.3758/MC.38.8.1087 [DOI] [PubMed] [Google Scholar]
  • 15.Swire B, Ecker UKH, Lewandowsky S. The role of familiarity in correcting inaccurate information. J Exp Psychol Learn Mem Cogn. 2017;43:1948–61. doi: 10.1037/xlm0000422 [DOI] [PubMed] [Google Scholar]
  • 16.Bower GH, Morrow DG. Mental models in narrative comprehension. Science. 1990;247:44–8. doi: 10.1126/science.2403694 [DOI] [PubMed] [Google Scholar]
  • 17.Glenberg A, Meyer M, Lindem K. Mental models contribute to foregrounding during text comprehension. J Mem Lang. 1987;26(1):69–83. 10.1016/0749-596X(87)90063-5 [DOI] [Google Scholar]
  • 18.Johnson-Laird PN. Mental models and probabilistic thinking. Cognition. 1994;50:189–209. doi: 10.1016/0010-0277(94)90028-0 [DOI] [PubMed] [Google Scholar]
  • 19.Brydges CR, Gignac GE, Ecker UKH. Working memory capacity, short-term memory capacity, and the continued influence effect: A latent-variable analysis. Intelligence. 2018;69:117–22. 10.1016/j.intell.2018.03.009 [DOI] [Google Scholar]
  • 20.Ecker UKH, Lewandowsky S, Swire B, Chang D. Correcting false information in memory: Manipulating the strength of misinformation encoding and its retraction. Psychon Bull Rev. 2011;18(3):570–8. doi: 10.3758/s13423-011-0065-1 [DOI] [PubMed] [Google Scholar]
  • 21.Kendeou P, Butterfuss R, Kim J, Van Boekel M. Knowledge revision through the lenses of the three-pronged approach. Mem Cognit. 2019;47:33–46. doi: 10.3758/s13421-018-0848-y [DOI] [PubMed] [Google Scholar]
  • 22.Kendeou P, Walsh EK, Smith ER, O’Brien EJ. Knowledge revision processes in refutation texts. Discourse Process. 2014;51:374–97. 10.1080/0163853X.2014.913961 [DOI] [Google Scholar]
  • 23.Gordon A, Brooks JCW, Quadflieg S, Ecker UKH, Lewandowsky S. Exploring the neural substrates of misinformation processing. Neuropsychologia. 2017;106:216–24. doi: 10.1016/j.neuropsychologia.2017.10.003 [DOI] [PubMed] [Google Scholar]
  • 24.Gordon A, Quadflieg S, Brooks JC, Ecker UK, Lewandowsky S. Keeping track of ‘alternative facts’: The neural correlates of processing misinformation corrections. Neuroimage. 2019;193:46–56. doi: 10.1016/j.neuroimage.2019.03.014 [DOI] [PubMed] [Google Scholar]
  • 25.Gernsbacher MA. Surface information loss in comprehension. Cogn Psychol. 1985;17(3):324–63. doi: 10.1016/0010-0285(85)90012-X [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Radvansky GA. Across the event horizon. Curr Dir Psychol Sci. 2012;21(4):269–72. 10.1177/0963721412451274 [DOI] [Google Scholar]
  • 27.Zwaan RA, Radvansky GA. Situation models in language comprehension and memory. Psychol Bull. 1998;123(2):162–85. doi: 10.1037/0033-2909.123.2.162 [DOI] [PubMed] [Google Scholar]
  • 28.Kurby CA, Zacks JM. Segmentation in the perception and memory of events. Trends Cogn Sci. 2008;12(2):72–9. doi: 10.1016/j.tics.2007.11.004 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Zacks JM, Speer NK, Swallow KM, Braver TS, Reynolds JR. Event perception: a mind-brain perspective. Psychol Bull. 2007;133(2):273–93. doi: 10.1037/0033-2909.133.2.273 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Zacks JM, Swallow K. Event segmentation. Curr Dir Psychol Sci. 2007;16(2):80–4. doi: 10.1111/j.1467-8721.2007.00480.x [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Bauer PJ, Varga NL. Similarity and deviation in event segmentation and memory integration: Commentary on Richmond, Gold, & Zacks. J Appl Res Mem Cogn. 2017;6(2):124–8. 10.1016/j.jarmac.2017.01.006 [DOI] [Google Scholar]
  • 32.Magliano JP, Miller J, Zwaan RA. Indexing space and time in film understanding. Appl Cogn Psychol. 2001;15(5):533–45. 10.1002/acp.724 [DOI] [Google Scholar]
  • 33.Pettijohn KA, Radvansky GA. Narrative event boundaries, reading times, and expectation. Mem Cognit. 2016;44(7):1064–75. doi: 10.3758/s13421-016-0619-6 [DOI] [PubMed] [Google Scholar]
  • 34.Richmond LL, Zacks JM. Constructing experience: Event models from perception to action. Trends Cogn Sci. 2017;21(12):962–80. doi: 10.1016/j.tics.2017.08.005 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Zacks JM, Speer NK, Reynolds JR. Segmentation in reading and film comprehension. J Exp Psychol Gen. 2009;138(2):307–27. doi: 10.1037/a0015305 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Kurby CA, Zacks JM. Starting from scratch and building brick by brick in comprehension. Mem Cognit. 2012;40(5):812–26. doi: 10.3758/s13421-011-0179-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Swallow KM, Zacks JM, Abrams R. Event Boundaries in Perception Affect Memory Encoding and Updating. J Exp Psychol Gen. 2009;138(2):236–57. doi: 10.1037/a0015631 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Horner AJ, Bisby JA, Wang A, Bogus K, Burgess N. The role of spatial boundaries in shaping long-term event representations. Cognition. 2016;154:151–64. doi: 10.1016/j.cognition.2016.05.013 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Radvansky GA, Copeland DE. Walking through doorways causes forgetting: Situation models and experienced space. Mem Cognit. 2006;34(5):1150–6. doi: 10.3758/bf03193261 [DOI] [PubMed] [Google Scholar]
  • 40.Radvansky GA, Copeland DE. Reading times and the detection of event shift processing. J Exp Psychol Learn Mem Cogn. 2010;36(1):210–6. doi: 10.1037/a0017258 [DOI] [PubMed] [Google Scholar]
  • 41.Rinck M, Bower GH. Temporal and Spatial Distance in Situation Models. Mem Cognit. 2000;28(8):1310–20. doi: 10.3758/bf03211832 [DOI] [PubMed] [Google Scholar]
  • 42.Rinck M, Weber U. Who when where: An experimental test of the event-indexing model. Mem Cognit. 2003;31(8):1284–92. doi: 10.3758/bf03195811 [DOI] [PubMed] [Google Scholar]
  • 43.Swallow KM, Barch DM, Head D, Maley CJ, Holder D, Zacks JM. Changes in events alter how people remember recent information. J Cogn Neurosci. 2011;23(5):1052–64. doi: 10.1162/jocn.2010.21524 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Zwaan RA. Processing narrative time-shifts. J Exp Psychol Learn Mem Cogn. 1996;22(5):1196–207. 10.1037/0278-7393.22.5.1196 [DOI] [Google Scholar]
  • 45.Zwaan RA, Langston M, Graesser A. The Construction of Situation Models in Narrative Comprehension: An Event-Indexing Model. Psychol Sci. 1995;6(5):292–7. 10.1111/j.1467-9280.1995.tb00513.x [DOI] [Google Scholar]
  • 46.Zwaan RA, Magliano JP, Graesser AC. Dimensions of situation model construction in narrative comprehension. J Exp Psychol Learn Mem Cogn. 1995;21(2):386–97. 10.1037/0278-7393.21.2.386 [DOI] [Google Scholar]
  • 47.Brown GD, Neath I, Chater N. A temporal ratio model of memory. Psychol Rev. 2007;114(3):539–76. doi: 10.1037/0033-295X.114.3.539 [DOI] [PubMed] [Google Scholar]
  • 48.Speer NK, Zacks JM, Reynolds JR. Human brain activity time-locked to narrative event boundaries. Psychol Sci. 2007;18(5):449–55. doi: 10.1111/j.1467-9280.2007.01920.x [DOI] [PubMed] [Google Scholar]
  • 49.Robin J, Buchsbaum BR, Moscovitch M. The primacy of spatial context in the neural representation of events. J Neurosci. 2018:1–41. doi: 10.1523/JNEUROSCI.1638-17.2018 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.Smith SM, Vela E. Environmental context-dependent memory: A review and meta-analysis. Psychon Bull Rev. 2001;8(2):203–20. doi: 10.3758/bf03196157 [DOI] [PubMed] [Google Scholar]
  • 51.Radvansky GA, Krawietz SA, Tamplin AK. Walking through doorways causes forgetting: Further explorations. Q J Exp Psychol. 2011;64(8):1632–45. doi: 10.1080/17470218.2011.571267 [DOI] [PubMed] [Google Scholar]
  • 52.Radvansky GA, Tamplin AK, Krawietz SA. Walking through doorways causes forgetting: Environmental integration. Psychon Bull Rev. 2010;17(6):900–4. doi: 10.3758/PBR.17.6.900 [DOI] [PubMed] [Google Scholar]
  • 53.Lawrence Z, Peterson D. Mentally walking through doorways causes forgetting: The location updating effect and imagination. Memory. 2016;24(1):12–20. doi: 10.1080/09658211.2014.980429 [DOI] [PubMed] [Google Scholar]
  • 54.Pettijohn KA, Radvansky GA. Walking through doorways causes forgetting: environmental effects. J Cogn Psychol. 2016;28(3):329–40. 10.1080/20445911.2015.1123712 [DOI] [Google Scholar]
  • 55.Pettijohn KA, Radvansky GA. Walking through doorways causes forgetting: recall. Memory. 2018;26(10):1430–5. doi: 10.1080/09658211.2018.1489555 [DOI] [PubMed] [Google Scholar]
  • 56.Radvansky GA, Pettijohn KA, Kim J. Walking through doorways causes forgetting: Younger and older adults. Psychol Aging. 2015;30(2):259–65. doi: 10.1037/a0039259 [DOI] [PubMed] [Google Scholar]
  • 57.Godden DR, Baddeley AD. Context‐dependent memory in two natural environments: On land and underwater. Br J Psychol. 1975;66(3):325–31. 10.1111/j.2044-8295.1975.tb01468.x [DOI] [Google Scholar]
  • 58.Thomson DM, Tulving E. Associative encoding and retrieval: Weak and strong cues. J Exp Psychol. 1970;86(2):255–62. doi: 10.1037/h0029997 [DOI] [Google Scholar]
  • 59.Wahlheim CN, Alexander TT, Peske CD. Reminders of Everyday Misinformation Statements Can Enhance Memory for and Beliefs in Corrections of Those Statements in the Short Term. Psychol Sci. 2020;31(10):1325–39. doi: 10.1177/0956797620952797 [DOI] [PubMed] [Google Scholar]
  • 60.Wahlheim CN, Zacks JM. Memory Guides the Processing of Event Changes for Older and Younger Adults. J Exp Psychol Gen. 2019;148(1):30–50. doi: 10.1037/xge0000458 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 61.Cohn-Sheehy BI, Delarazan AI, Reagh ZM, Kim K, Barnett AJ, Zacks JM, et al. The hippocampus constructs narrative memories across distant events. Current Biology. 2021;31:4935–45. doi: 10.1016/j.cub.2021.09.013 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 62.Ecker UKH, Hogan JL, Lewandowsky S. Reminders and repetition of misinformation: Helping or hindering its retraction? J Appl Res Mem Cogn. 2017:1–8. 10.1016/j.jarmac.2017.01.014 [DOI] [Google Scholar]
  • 63.Ecker UKH, O’Reilly Z, Reid JS, Chang EP. The effectiveness of short‐format refutational fact‐checks. Br J Psychol. 2020;111(1):36–54. doi: 10.1111/bjop.12383 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 64.Faul F, Erdfelder E, Lang AG, Buchner A. G* Power 3: A flexible statistical power analysis program for the social, behavioural, and biomedical sciences. Behav Res Methods. 2007;39(2):175–91. doi: 10.3758/bf03193146 [DOI] [PubMed] [Google Scholar]
  • 65.Brysbaert M. How Many Participants Do We Have to Include in Properly Powered Experiments? A Tutorial of Power Analysis with Reference Tables. J Cogn. 2019;2(1):1–38. doi: 10.5334/joc.54 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 66.Lakens D. Performing high‐powered studies efficiently with sequential analyses. Eur J Soc Psychol. 2014;44(7):701–10. 10.1002/ejsp.2023 [DOI] [Google Scholar]
  • 67.Schott E, Rhemtulla M, Byers-Heinlein K. Should I test more babies? Solutions for transparent data peeking. Infant Behav Dev. 2019;54:166–76. doi: 10.1016/j.infbeh.2018.09.010 [DOI] [PubMed] [Google Scholar]
  • 68.Litman L, Robinson J, Abberbock T. TurkPrime.com: A versatile crowdsourcing data acquisition platform for the behavioral sciences. Behav Res Methods. 2016:1–10. doi: 10.3758/s13428-016-0727-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 69.Wagenmakers EJ, Wetzels R, Borsboom D, Van Der Maas HL. Why psychologists must change the way they analyze their data: the case of psi: comment on Bem (2011). J Pers Soc Psychol. 2011;100:426–32. doi: 10.1037/a0022790 [DOI] [PubMed] [Google Scholar]
  • 70.Radvansky GA, Tamplin AK, Armendarez J, Thompson AN. Different kinds of causality in event cognition. Discourse Process. 2014;51(7):601–18. 10.1080/0163853X.2014.903366 [DOI] [Google Scholar]
  • 71.Radvansky GA, Zacks JM. Event perception. Wiley Interdiscip Rev Cogn Sci. 2011;2(6):608–20. doi: 10.1002/wcs.133 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 72.Radvansky GA, Zacks RT. The retrieval of situation-specific information. In: Conway MA, editor. Cognitive models of memory: Mit Press; 1997. p. 173–213. [Google Scholar]
  • 73.Langston W, Kramer DC, Glenberg AM. The representation of space in mental models derived from text. Mem Cognit. 1998;26(2):247–62. doi: 10.3758/bf03201137 [DOI] [PubMed] [Google Scholar]
  • 74.Radvansky GA, Copeland DE. Functionality and spatial relations in memory and language. Mem Cognit. 2000;28(6):987–92. doi: 10.3758/bf03209346 [DOI] [PubMed] [Google Scholar]
  • 75.Therriault DJ, Rinck M, Zwaan RA. Assessing the influence of dimensional focus during situation model construction. Mem Cognit. 2006;34(1):78–89. doi: 10.3758/bf03193388 [DOI] [PubMed] [Google Scholar]
  • 76.Wilson SG, Rinck M, McNamara TP, Bower GH, Morrow DG. Mental models and narrative comprehension: Some qualifications. J Mem Lang. 1993;32(2):141–54. 10.3758/BF03202594 [DOI] [Google Scholar]
  • 77.Zwaan RA, van Oostendorp H. Do readers construct spatial representations in naturalistic story comprehension? Discourse Process. 1993;16(1–2):125–43. 10.1080/01638539309544832 [DOI] [Google Scholar]
  • 78.Bailey HR, Kurby CA, Sargent JQ, Zacks JM. Attentional focus affects how events are segmented and updated in narrative reading. Mem Cognit. 2017;45(6):940–55. doi: 10.3758/s13421-017-0707-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 79.Thompson AN, Radvansky GA. Event boundaries and anaphoric reference. Psychon Bull Rev. 2016;23(3):849–56. doi: 10.3758/s13423-015-0961-x [DOI] [PubMed] [Google Scholar]
  • 80.Ecker UKH, Rodricks AE. Do False Allegations Persist? Retracted Misinformation Does Not Continue to Influence Explicit Person Impressions. J Appl Res Mem Cogn. 2020;9(4):587–601. [Google Scholar]
  • 81.Sanderson JA, Gignac GE, Ecker UKH. Working memory capacity, removal efficiency and event specific memory as predictors of misinformation reliance. J Cogn Psychol. 2021;33(5):518–32. 10.1080/20445911.2021.1931243 [DOI] [Google Scholar]
  • 82.Rinck M, Bower GH. Anaphora resolution and the focus of attention in situation models. J Mem Lang. 1995;34(1):110–31. 10.1006/jmla.1995.1006 [DOI] [Google Scholar]
  • 83.Pettijohn KA, Thompson AN, Tamplin AK, Krawietz SA, Radvansky GA. Event boundaries and memory improvement. Cognition. 2016;148:136–44. doi: 10.1016/j.cognition.2015.12.013 [DOI] [PubMed] [Google Scholar]
  • 84.Sahakyan L, Hendricks HE. Context change and retrieval difficulty in the list-before-last paradigm. Mem Cognit. 2012;40(6):844–60. doi: 10.3758/s13421-012-0198-0 [DOI] [PubMed] [Google Scholar]
  • 85.Dai Y, Yu W., Shen F. The effects of message order and debiasing information in misinformation correction. Int J Commun. 2021;15:1039–59. [Google Scholar]
  • 86.Ecker UKH, Lewandowsky S, Cheung CSC, Maybery MT. He did it! She did it! No, she did not! Multiple causal explanations and the continued influence of misinformation. J Mem Lang. 2015;85:101–15. 10.1016/j.jml.2015.09.002 [DOI] [Google Scholar]
  • 87.Swire-Thompson B, Cook J, Butler LH, Sanderson JA, Lewandowsky S, Ecker UKH. Evidence for a limited role of correction format when debunking misinformation. CR:PI. 2021;6(83):1–15. doi: 10.1186/s41235-021-00346-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 88.Baddeley AD, Hitch GJ. The recency effect: Implicit learning with explicit retrieval? Mem Cognit. 1993;21:146–55. doi: 10.3758/bf03202726 [DOI] [PubMed] [Google Scholar]
  • 89.Davelaar EJ, Goshen-Gottstein Y, Ashkenazi A, Haarmann HJ, Usher M. The demise of short-term memory revisited: empirical and computational investigations of recency effects. Psychol Rev. 2005;112:3–42. doi: 10.1037/0033-295X.112.1.3 [DOI] [PubMed] [Google Scholar]
  • 90.Wilkes AL, Leatherbarrow M. Editing episodic memory following the identification of error. Q J Exp Psychol. 1988;40(2):361–87. 10.1080/02724988843000168 [DOI] [Google Scholar]
  • 91.Brydges CR, Gordon A, Ecker UKH. Electrophysiological correlates of the continued influence effect of misinformation: an exploratory study. J Cogn Psychol. 2020;32(8):771–84. 10.1080/20445911.2020.1849226 [DOI] [Google Scholar]
  • 92.Sanderson JA, Farrell S, Ecker UKH. Examining the Reproducibility of Temporal Event Boundaries in Continued Influence Effect and Event Segmentation Paradigms. Psyarxiv [preprint] 2022:1–84. Available from: https://psyarxiv.com/6zn2a [Google Scholar]

Decision Letter 0

Shane Lindsay

1 Mar 2022

PONE-D-21-32228Examining the role of information integration in the continued influence effect using an event segmentation approachPLOS ONE

Dear Dr. Sanderson,

Thank you for submitting your manuscript to PLOS ONE. I have now received comments on your manuscript from two expert reviewers and I have also studied your manuscript myself. Both reviews were positive, with comments that were constructive and clear. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

A couple of points stood out for further comment. I agree with Reviewer 2 that results in the supplementary materials section would be better off in the main results section.

More substantively, like Reviewer 1, I also wondered how the effect used in the power analyses of Experiment 1 was chosen. Furthermore, these f effect sizes can be difficult to understand so I wonder if it can be also explained in more concrete terms (see below, and in relation to the specific test(s) used). I highlight this in particular as in the face of null results an important question to ask is how likely was a type 2 error to have occured. While the Bayes Factors did give good support for the null hypotheses, the sample sizes used here for a between-subjects design will likely have limited sensitivity to detect small effect sizes. PLOS-ONE is a journal where our policy is to accept null results where the methods are sound, but in these cases I would normally like to see more consideration of statistical power and discussion of the sensitivity of the design to detect an effect if one were present.

As I understand it, in the case here I note that the main statistical hypothesis tested in the results of Experiment 1 was a difference in conditions in the size of the inference effect - a main effect in a Bayesian ANOVA, with three conditions. However, the hypotheses in the introduction did not straightforwardly map onto this test, with the following hypotheses:

H1: "both no-shift and shift conditions would reduce reliance on misinformation compared to a no-retraction control"

H2: "integration would be facilitated in the no shift condition... relative to the shift condition"

The exact of test of H1 was unclear to me - would this be the average of the no-shift + shift conditions compared with the control condition? H2 seemed to be more in line with a t-test (as in Experiment 2). I suspect that the study had good power to detect a CIE but may have limited power to detect a modulation of the CIE unless the modulation was large. Relating back to the power analysis used as a justification, it mentions "detecting an effect", but not which one, as there are several on offer here (i.e. the CIE effect and effects associated with the 2 hypotheses). This could be clarified, with more detail on the mapping of your hypotheses to tests. Furthermore, with frequentist tests it is straightforward and informative to conduct a sensitivity analyses of your design - this is not conditional on the effect you observed, but what effect sizes you would have good power to detect given the properties of your design and test used. Given the use of Bayesian analysis this might be less straightforward to conduct, but as mentioned above in the face of a null result greater consideration of sensitivity seems necessary, especially as in the discussion it makes reference to numerical differences in the effects across conditions.

Please submit your revised manuscript by Apr 15 2022 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

We look forward to receiving your revised manuscript.

Kind regards,

Shane Lindsay

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at 

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. We note that the grant information you provided in the ‘Funding Information’ and ‘Financial Disclosure’ sections do not match. 

When you resubmit, please ensure that you provide the correct grant numbers for the awards you received for your study in the ‘Funding Information’ section.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Partly

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: Reasoning about events can be negatively affect by earlier misinformation even after that information has been corrected. This so-called continued influence effect has been attributed to failures to recollection corrections and to integrative encoding of correct and incorrect information. The present study examined the role of integrative encoding in this effect in the context of event segmentation and memory updating. According to a prominent theory of event comprehension, when people experience event boundaries during misinformation correct, integrative encoding and, thus, memory updating should be disrupted, leading to a greater negative influence of misinformation on subsequent memory accuracy. The presence of spatial event boundaries was manipulated during retractions to examine differences in integrative encoding effects on memory updating. The results from two experiments were mixed as event boundaries had both no effect on memory for corrections (Experiment 1) and reduced the negative effects of misinformation on subsequent memory accuracy (Experiment 2). These findings do not support an integrative encoding account of the continued influence effect and better align with the view that this effect occurs due to retrieval failure.

The topic of the mechanisms underlying the negative proactive effects of misinformation on memory for correct information is quite timely given the high prevalence of misinformation in everyday life. The present study takes a unique approach to illuminating the role of various encoding and retrieval mechanisms in the updating of memories for inaccurate information. I appreciated the creative connection between event segmentation theory and the continued influence effect as well as the innovative experimental procedure. However, the study had some key limitations, such as the ambiguity in results across experiments. On balance, this package seems appropriate for the journal with the caveat that the inconsistency in results across experiments should be highlighted as an impediment for clear interpretation. Below, I describe some concerns and offer specific comments that I hope will help improve the manuscript.

One general issue was that the Introduction did not summary a relevant literature on the role of retrieval during encoding in information integration. The present study appears to examine how disruptions in the contact between misinformation and subsequent retractions impacts memory updating. There are studies from the memory and discourse comprehension literatures that should inspire predictions in the current experiments. Specifically, there is recent memory work from Wahlheim and Zacks (2019), Wahlheim et al. (2020), and Cohen-Sheehy et al. (2021) on this topic. There is also earlier discourse comprehension work from Kintsch and more recent work from Kendeou and colleagues that may be relevant here. The authors may consider how the notion of integrative encoding (information integration) in those studies fits in here. In my view, the findings from these studies showing that memory updating is more effective when recent information elicits retrieval of earlier information also leads to the prediction that reducing the accessibility of earlier information via event boundaries should lead to fewer retrievals and therefore less integrative encoding. This perspective could provide a more complete mechanistic motivation for the hypotheses given in the present study.

Specific Comments:

p. 5, line 99: Do event boundaries decrease the “availability” or “accessibility” of information in memory?

p. 7: It would be great if there was a section here that synthesized the literature review and provided an overarching description of “The Present Study” to give the reader a better understanding of why the three literatures in the Introduction were considered.

pp. 7 and 8, lines 147-152: What was the rationale for using a between-subjects design instead of a within-subjects design? Was there any concern about sampling from multiple populations for the experimental (undergraduates) and control (MTurk workers) conditions? More generally, I could not understand from the description of the experiment what the control condition was intended to assess with respect to the main hypotheses.

p. 8, Participants: How did the authors choose the effect size that was targeted in the power analysis?

p. 9, Event reports: It would be great to have more information about what motivated the structure of the retractions. Specifically, the example here does not include details from the original misinformation statement. Was this done purposefully to decrease the contact between retractions and misinformation to increase the chances that the event boundary manipulation would reduce the accessibility of misinformation? I have a similar query about the materials in Experiment 2.

p. 10, lines 196-201: Do the authors worry that providing response alternatives on the questionnaire had reactive effects on subsequent memory for event details? I have a similar query about the materials in Experiment 2.

p. 10, line 220: I believe the monitor was a 21.5-inch (not 31.5 inch).

p. 11, Procedure: What motivated the decision to administer the final memory test in a unique room (i.e., Room C)? I had the impression that the effects of spatial shift would be most pronounced if the final testing room context matched the retraction encoding context.

p. 12, Memory scores: Are the authors concerned that providing response alternatives on a multiple-choice memory test rescues impaired event details accessibility post-boundary? In other words, could available but less accessible details be made more accessible with the presentation of copy cues during the test?

p. 13, lines 287-290: I was uncertain how the comparison presented here provided support for a continued influence effect. Shouldn’t that require a comparison of misinformation correction items with control items for which no misinformation had appeared? If I have just missed something, it would be great if the authors could explain that here.

p. 14, Discussion: Echoing a concern above, I am not confident that the comparison of experimental groups from one population should be compared directly with a control group from another population. Also, for the possible explanations as to why Experiment 1 failed, the authors may consider my point above that a recognition task could rescue what would otherwise be inaccessible memories of event details following spatial shifts. The authors allude to this possibility in lines 328-329.

pp. 21 and 22: I was confused about the conversion of charge scores into 0 and 10 values. Could the authors have treated this as a 1/0 variable and used some type of logistic regression approach?

General Discussion: This section is a bit shorter than I am accustomed to reading in multi-experiment reports. The authors might consider connecting their findings to the literature more broadly here.

Reviewer #2: In two behavioral studies, Sanderson and colleagues utilized an event segmentation approach to investigate the persistence of misinformation. Given the robustness of the continued influence effect, finding new ways to investigate the phenomenon is a very worthwhile pursuit. Overall, I find the procedure and results easy to follow, and I think the VR detective framing (Experiment 2) is quite clever. Below are a few comments and suggestions which may help further strengthen the paper.

Introduction:

1. As I progressed through the introduction, I was anticipating a study that manipulated event segmentation within the context of the narratives. As such, I was surprised when I came to find out that the manipulation was segmentation of the encoding environment. Further motivation and clarification of that manipulation choice in the introduction will be important.

I should add that although I am generally familiar with the event segmentation literature, I was unfamiliar with the “location updating effect” described on page 6. When I read the description of “…when an agent moves from one location to another…”, I was assuming that the “agent” refers to the protagonist of the narrative. Clarifying that section -- and further explaining how this goes beyond the traditional context-dependent memory effects -- will go a long way towards minimizing potential confusion.

Experiment 1:

2. In coding the open-ended inference question, the authors opted for a broad definition of “no misinformation reliance”, page 12: “If participants recalled the retraction and/or the alternative, they received a score of 0 (indicating no misinformation reliance) irrespective of whether they also recalled the critical information…”

Recent work has shown that there may be an important distinction between different operationalizations of “continued influence” (https://doi.org/10.1186/s41235-021-00335-9). If the open-ended responses were coded differently -- for example, 0 when they recalled only the retraction and/or alternative; 5 points when they recalled retraction and/or alternative AND critical information; 10 = critical information only -- will the patterns look different? This may be a more nuanced way to distinguish among the varying levels of success at integration. This maybe an interesting exploratory analysis, especially since the initial analyses did not yield any effect of segmentation conditions.

3. Minor point: I assume the researchers converted the open-ended question score to 0 or 10 so that the scale would match the close-ended questions? It would be helpful to include an explicit statement about that in the scoring section.

Experiment 2:

4. In Part 2 of the report, retractions of non-critical information were introduced alongside retractions of critical information. Please clarify the rationale for and utility of this design choice.

5. Unless dictated by a prescribed word limit, I strongly encourage the authors to incorporate the Results reported in the Supplement into the main text.

General Discussion:

6. On page 24, the authors wrote, “Both experiments found that a retraction was effective in reducing reliance on misinformation …” Since Experiment 2 did not include a baseline (no retraction) condition, this claim does not apply to Experiment 2. Please rephrase.

7. The paper will benefit from some discussion of the future utility of using the event segmentation approach as a way to understand CIE.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

Decision Letter 1

Shane Lindsay

19 May 2022

PONE-D-21-32228R1

Examining the role of information integration in the continued influence effect using an event segmentation approach

PLOS ONE

Dear Dr. Sanderson,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

I was not able to secure the two previous reviewers to review the re-submission. To expedite the review process, I have considered the revised submission and your responses, and have the following comments below. These are focused on analyses and corresponding conclusions in response to revisions or outstanding issues from the original submission. 

Please submit your revised manuscript by Jul 02 2022 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Shane Lindsay

Academic Editor

PLOS ONE

Journal Requirements:

Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

Additional Editor Comments:

1. Myself and Reviewer 1 asked for a justification for the sample size. This was provided in the cover letter but not in the revisions. Could the rationale be provided in the manuscript. If mentioning a previous finding as a justification, could you provide more detail (e.g., exactly which result/test in the Rinck and Bower paper is being used as justification).

2. Further in regard to the justification, this appears based on a frequentist power analyses with a p < .05 decision threshold, which does not match the inference procedure and tests used in the paper. Could the rationale of this approach be described in the paper.

3. "All Bayesian analyses were conducted in JASP, using default priors.". There was previously some text which described the priors but this was removed. A reproducible analysis script is not available and the JASP version was not given. The default priors in JASP could be subject to change, and hence reproducibility of the analyses would be compromised. Therefore could some information on the priors be provided in the manuscript.

4. "A Bayesian one-way between-subjects ANOVA confirmed there were no differences between no-shift and shift conditions, F(1, 108) = .02, BF01 = 4.91." The word significant was removed in revisions. I understand the reason for this change since a NHST test was not performed. However, there was a numerical difference so this statement is incorrect. Similar comments apply to the results on p. 25 and perhaps elsewhere, e.g., "a Bayesian independent-samples t-test confirmed there was no difference between conditions, t(98) = .28, BF01 = 4.58". The description of the analysis on p.14 describes ranges of Bayes Factors for which qualitative labels for evidence are assigned based on conventions established in the literature. However, the inferential process that leads to the conclusion of "no difference" based on a BF value is not described, and the pre-registration does not articulate this either. In frequentist analysis the phrase "statistically significant difference" is used to qualify such statements, but such terminology to indicate qualification appropriate to Bayesian hypothesis testing is (mostly) absent here. Throughout, could results be described in a way that it is more clear as to what inference is being performed (with explanation of the inference procedure earlier on if necessary), and to qualify statements the results, i.e., it would seem more appropriate to talk about evidence for effects, or more how the Bayes Factor favours one model over another, rather than categorical statements of a presence/absence of an effect.

5. I apologise but I am still confused on the use of the ANOVA. On page 15 where this ANOVA is described, the means for three conditions were provided (i.e. including the control), but the following statement of a description of just a difference between two conditions (no-shift and shift), which would result from an ANOVA with two levels or a t-test. The following analysis on page 16 again described means for the control conditions along with shift and no-shift conditions, but here t-tests were used. My confusion here is that it does not appear that the control condition was included in any statistical analyses, and therefore I was unclear on the purpose of having it. Could this be explained in the text.

6. Page 26 reports Bayesian chi-square tests. There is not sufficient detail to allow reproducibility of these analyses in the manuscript. Could more explanation be provided, including priors used.

7. I did wonder why the approach taken in Experiment 2 was not used in Experiment 1 - an effect was found in the opposite direction in both cases but only Experiment 2 was a one-tailed test performed. I presume that the reason was that that in Experiment 2 there was weak evidence for the alternative hypothesis, and moderate evidence for the null in Experiment 1. But I wondered how the results of Experiment 1 would change (i.e. make the null findings more equivocal). I wonder if this was considered, and if it might be helpful, but I am not suggesting this is necessary.

     8. Sensitivity to small effect sizes:

  • Previously I asked for more consideration of design to detect small effect sizes. The response letter stated:

  • "We have conducted a sensitivity analysis (using G*Power) for an independent samples t-test which found that our sample was large enough to detect an effect size of f = .27."

  • The purpose of a sensitivity analysis would be to see the range of possible effect sizes that your design had sensitivity to detect. That you had a good power (in frequentist terms) to detect medium to large effects (i.e. f > .2) was not the issue being raised. A further sensitivity analysis was provided in the response editor letter compares H0 with f = 0 with H1 with f = 2 (d = .4). This is comparing a null effect with a medium-ish effect. The point I wanted to see addressed or discussed was the sensitivity of your design to smaller effects.

  • "Additionally, in Experiment 1 the numeric effect between no-shift and shift conditions was in the opposite direction to our hypothesis, which speaks against a false-negative." 

  • Three possibilities can be considered to exist here. There is a true effect in the opposite direction to that predicted, a true null effect, or an true effect in the predicted direction. While it would be unlikely to find a Type S error (i.e., an error of the wrong sign) with a large effect in the predicted direction, the probability of this occurring with a smaller effect is non-negligible. The balance of evidence may favour one (or two) of these accounts over the other, but there does not seem to be very strong evidence to rule out any of these possibilities.

  • Note, in Experiment 1, the difference in conditions appeared to be around d = .2, based on the difference in the means and an approximate calculation of the SD. Like Experiment 2, the effect was in the opposite direction as predicted. The BF01 was 3.02. A Bayes Factor provides a continuous measure, but this value is very close to the 3 threshold by some conventions the evidence could be described as anecdotal. Although numerical differences are highlighted in the text, the conclusions sometimes steer towards a more categorical interpretation, e.g. wordings such as "no effect, e.g. " no measurable impact", "failed to find an effect". Yet, with a frequentist analysis, the power would be approximately .18 for a similar sized effect (i.e., in r power.t.test(n=54, d=.2, type="two.sample"). The actual power is unknown, but my point is that a Bayesian analysis cannot overcome the same issue that a frequentist analysis has with detecting small effects with a limited sample size - in Bayesian terms, the probability of finding convincing evidence or misleading evidence (based on some threshold of evidence, e.g.  Schönbrodt & Wagenmakers, 2018).

  • It seems to me that the evidence seems pretty convincing that there is not a medium or large effect, but the evidence for ruling out a small effect is not convincing. And based on the results of Experiment 2, a small effect in the opposite direction as predicted in Experiment 1 does seem possible. I do note that that in terms of your conclusions, either a null effect or an effect in the opposite direction in Experiment 1 does not change your substantive conclusions, hence this may be considered a minor point, and it is quite a generic one - i.e. most psychology designs use sample sizes in a range that make it difficult to detect small effects. Consequently, I don't think a sensitivity analysis is necessary in a revision, but this point can easily be addressed by more judicious use of language to describe evidence for effects in the results (as mentioned in comments above) and corresponding language to match the description of results in the discussion, and what can and cannot be plausibly ruled out.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

Decision Letter 2

Shane Lindsay

4 Jul 2022

Examining the role of information integration in the continued influence effect using an event segmentation approach

PONE-D-21-32228R2

Dear Dr. Sanderson,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Shane Lindsay

Academic Editor

PLOS ONE

Acceptance letter

Shane Lindsay

8 Jul 2022

PONE-D-21-32228R2

Examining the role of information integration in the continued influence effect using an event segmentation approach

Dear Dr. Sanderson:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Shane Lindsay

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 File. Supplemental information for Experiments 1 and 2.

    (DOCX)

    Attachment

    Submitted filename: Response to Reviewers.docx

    Attachment

    Submitted filename: Response to Reviewers.docx

    Data Availability Statement

    Materials are provided in the Supplement. Data are available on the Open Science Framework website at https://osf.io/qrpjb/ (doi:10.17605/OSF.IO/QRPJB; Experiment 1) and https://osf.io/sxh2w/ (doi:10.17605/OSF.IO/SXH2W; Experiment 2).


    Articles from PLoS ONE are provided here courtesy of PLOS

    RESOURCES