Skip to main content
PLOS One logoLink to PLOS One
. 2020 Mar 6;15(3):e0229197. doi: 10.1371/journal.pone.0229197

Playing with fire. Understanding how experiencing a fire in an immersive virtual environment affects prevention behavior

Patty C P Jansen 1,2,¤,*, Chris C P Snijders 2, Martijn C Willemsen 2
Editor: Geilson Lima Santana3
PMCID: PMC7059903  PMID: 32142518

Abstract

A potentially effective way to influence people’s fire prevention behavior is letting them experience a fire in an immersive virtual environment (IVE). We analyze the effects of experiencing a fire in an IVE (versus an information sheet) on psychological determinants of behavior—knowledge, vulnerability, severity, self-efficacy, and locus of control—based mainly on arguments from Protection Motivation Theory and the Health Belief Model. Crucial in our setup is that we also relate these determinants to actual prevention behavior. Results show that IVE has the hypothesized effects on vulnerability, severity, and self-efficacy, and an unexpected negative effect on knowledge. Only knowledge and vulnerability showed subsequent indirect effects on actual prevention behavior. There remains a direct positive effect of IVE on prevention behavior that cannot be explained by any of the determinants. Our results contradict the implicit assumption that an induced change in these psychological determinants by IVE, necessarily implies a change in behavior. A recommendation for research on the effects of IVE’s is, whenever possible, to study the actual target behavior as well.

Introduction

For the year 2014, the US fire Administration [1] reported 379,500 residential building fires of which 12,075 cases resulted in an injury and 2,765 cases were fatal. The most common cause of these residential building fires were cooking fires [1]. For the Netherlands similar findings apply: a grease fire was the most important cause of fatal fire incidents [2]. The use of prevention measures can help to reduce this number: while a smoke alarm helps to signal a grease fire on time, a fire blanket or a fire extinguisher enables people to extinguish a grease fire before it expands. Insurance companies and society at large would benefit if more people would own and apply preventive measures.

While mass communication is a typical way to influence people’s behavior, mass communication about risks tends to only affect the judgement of risks on a societal level and not on a personal level [3]. In contrast, research concerning natural hazard experiences has shown that personal risk experience can stimulate self-protective and coping behavior [46]. Personal experience can influence preventive behaviors, and can even have a larger effect on preventive behaviors than general communication messages because of an increase in susceptibility and worry [7]. A precondition is that the experience is more severe than expected, and that the specific preventive measures are perceived to be effective [7].

In the current study we use an Immersive Virtual Environment (IVE) to have participants experience the large impact of a virtual grease fire and to experience how a simple preventive measure such as a fire blanket can reduce the impact of the fire. Our study has three characteristics that, taken together, are an addition to the current IVE literature. First, while there are studies in other domains that study the effects of experiencing a hazard in an IVE (e.g. flood risk, aircraft evacuation, terror attack), our study is the first to study the effects of experiencing a virtual fire. Second, we consider the effects of an IVE on a full set of “psychological determinants” (knowledge, vulnerability, severity, self-efficacy, and locus of control) and analyze them all together. Whenever we refer to psychological determinants in the remainder of this paper we refer to this set of concepts that are considered drivers for prevention behavior. Third, and most importantly, we measure actual prevention behavior, and test to what extent psychological determinants have an effect on actual behavior. This paper is structured as follows: we first present a brief overview of the use of IVEs to simulate risks. We then present our theoretical framework and hypotheses. Then, we test these hypotheses in a single estimation with a Structural Equation Model. Finally, we discuss the results and their implications.

Risk simulation in an Immersive Virtual Environment

Simulated experiences can be an effective way to influence people, as people often react to virtual experiences as if they were real [8]. They also allow people to (better) observe the link between cause and effect, which in turn can positively influence peoples’ attitudes and behaviors [8]. Next to the persuasive benefits that might arise after the message has been delivered, using virtual reality tools can increase the attractiveness of getting the message across, as people may generally be not interested or motivated enough to search for information themselves.

Studies on the effects of simulated risk experiences in immersive virtual environments (IVE), also known as virtual reality (VR), are relatively recent, although the argument as to why they might work has been around considerably longer. In this paper we also consider virtual 3D environments (the precursors of IVE) as a type of IVE. With IVE “users are perceptually surrounded by and immersed in an artificially-created environment” ([9], p. 57). Often, the user’s position and orientation is tracked through a tracking system and the user experiences the virtual environment and his orientation (i.e. if the user’s head turns right, visual information of right side of the IVE is perceived) through a head mounted display. Physical movement in the IVE can be performed by making use of advanced tracking systems on the user’s body (e.g. special gloves), possibly combined with a motion platform, but also more traditional analog devices (e.g. a joystick) can be used. Currently well-known consumer tools that offer IVE through a head-mounted display are the Oculus Rift, Samsung Gear, and the HTC Vive.

A distinctive characteristic of IVE is its influence on a human’s sense of presence: “the subjective experience of being in one place or environment, even when one is physically situated in another” ([10], p. 225). This increased sense of presence in IVE can lead to an increase or decrease in emotions, for example an increase of fear when confronted with a virtual spider [11] or the increase of aggressive feelings in the case of playing a violent videogame [9]. This sense of presence makes, amongst other factors, an IVE also a suitable research tool to study human behavior in a simulated environment, for example to study human behavior while in interaction with fire [12]. Another potential IVE application within the domain of fire, because of the ability to influence emotions and behavior, is to use IVE to influence people’s fire prevention behavior.

Several studies have shown that interactive (immersive) virtual environments in which risks are simulated can influence knowledge, emotions, attitudes, and intentions. For example, Zaalberg and Midden [13] have shown that an interactive 3D environment showing a flood simulation can result in increased motivation to search for information, increased motivation to evacuate, and a (small) increase in the willingness to buy flood insurance (compared to non-interactive 2D environments). Furthermore, Chittaro and Zangrando [14] have shown the impact of emotional intensity on awareness of personal fire safety in a fire evacuation game in an IVE. Their results have shown that the highly emotional game (with more visual and audio feedback) produced more anxiety and positively affected participants’ attitudes towards the dangers of smoke compared to a more mellow game. Another study of Chittaro and Sioni [15] in which a terror attack was simulated showed that the interactive 3D environment had more impact on risk perceptions than the non-interactive 3D simulation. In the domain of airplane safety, Chittaro [1618] has shown that a 3D serious game increased the knowledge about safety procedures and feelings of self-efficacy, and made participants feel more “in control” when confronted with emergency landings. This result is suggested to have positive behavioral implications since both self-efficacy as well as safety locus of control have proven to be important predictors for the performance of safety behaviors [1618]. Also, an IVE can result in more knowledge retention: an IVE in which people experienced a plane crash resulted in more knowledge concerning safety procedures, compared to a safety instruction card, one week after the intervention [19].

While it is promising that previous studies have shown positive effects on knowledge, locus of control, and other psychological determinants, such studies more or less implicitly assume that these effects are indicative of a subsequent improvement in behavior. There are hardly any studies in the safety domain that follow through on the effect of IVE on psychological determinants. In fact, as far as we know, the effects of risk simulation in an IVE on behavior are hardly measured at all, with or without the measurement of psychological determinants. This is the main contribution of our study, applied to the domain of fire prevention.

We now discuss a conceptual model that considers how IVE (when compared to non-IVE delivery of information) affects individuals’ psychological determinants with respect to (fire) prevention, and through these, may lead to preventive behavior.

Theoretical framework and hypotheses

From literature we have identified five psychological determinants that have an important role in influencing individuals’ prevention behavior and that are likely to be influenced by experiencing a fire in an IVE: knowledge, severity, vulnerability, self-efficacy, and locus of control. Three of these psychological determinants are grounded in two popular theories that offer a framework for the explanation of prevention behavior: Protection Motivation Theory (PMT) and the Health Belief Model (HMB). The HBM, originally developed by Rosenstock [20] and later discussed and revised by many scholars [2123], was developed to explain health prevention behavior. PMT was originally developed to explain the effects of fear arousing communications, referred to as “fear appeals” [24], and later extended to a more general theory on behavioral change [25]. Both theories use similar psychological determinants to explain behavior and are commonly used to explain preventive behaviors in the areas of health- and environmental risk. The basic idea behind both theories is that prevention behavior is driven by the evaluation of the risk (perceived vulnerability and severity), and by an evaluation of the coping response (barriers and benefits of the behavior, self-efficacy). The underlying arguments are equally appropriate for the area of fire prevention behavior, since this contains both a risk element as advised coping responses to deal with the risk. Besides these variables from PMT and HBM, knowledge about the topic and locus of control are also considered variables that may be influenced by IVE and may themselves influence subsequent prevention behavior [17,19,26]. Knowledge about the risk and possible risk-mitigating prevention behaviors, enables people to make the right preventive decisions [27]. While locus of control refers to the extent someone believes a negative event is something within their control, which is related to more safe behaviors [2831].

The link between these theories and risk simulation in an IVE can easily be established. Research on the effects of IVE’s often focuses on exactly the psychological determinants that are considered in PMT and HBM: after the exposure to the IVE someone may feel more vulnerable or at risk of some event (vulnerability), or may assess the consequences of potential events differently than they did before (severity) [6,1517], or may have more confidence in their ability to perform a specific behavior or in their coping ability (self-efficacy) [15,16,3234]. While almost all IVE research focuses on such potential “mental effects”, in many cases the ultimate goal is to affect individuals’ behavior. However, many researchers have left the relation between psychological determinants and actual behavior untested and sometimes unmentioned, probably in part because it is often difficult, or even impossible, to test the effects of IVE exposure on behavior.

While some studies have demonstrated the relationship between psychological determinants and behaviors [3537], this relationship is not always straightforward [26] and might depend on the context such as the domain, the behavior under study etc. We argue it is appropriate or at least worthwhile to study psychological determinants and the target behavior(s) simultaneously whenever possible, to test whether a change in determinants really results in a change in behavior.

Based on the arguments as suggested in the literature, and specific to our case of fire prevention, we created a conceptual model that outlines the effects we expect to find (see Fig 1). Although these psychological determinants have been used before and are similar across IVE studies, it still depends on the type of risk under study which determinants are relevant, and in what direction they should be influenced. For example, it is hardly necessary to increase the perceived severity of an emergency landing as people already consider this to be very severe; people should be convinced about their survival possibilities through influencing their knowledge and adoption of the advised safety procedures [16]. On the other hand, in the case of a fire, people are more likely to underestimate the severity, so in this context the aim is to positively influence the perceived severity of a fire [27].

Fig 1. Hypothesized model of the effects of IVE exposure relative to the INFO condition.

Fig 1

Squares represent observed constructs and oval shapes represent latent constructs.

It is important to note another characteristic of our study: we consider all effects of IVE as compared to a control group. Since the traditional way to persuade people to invest in fire prevention measures is by giving them information in text, it is relevant to compare the IVE experience with an information sheet as a control condition (referred to as: INFO condition). We expect that compared to this INFO condition, the IVE has a stronger effect on the psychological determinants, which in turn results in an increased likelihood of performing prevention behaviors in the IVE condition (compared to the INFO condition), as laid out in the hypothesized model in Fig 1. We now discuss the argumentation behind the hypothesized model in more detail for the separate psychological determinants and their effect on prevention behavior.

Knowledge

Knowledge is considered to be a likely determinant for performing the advised behaviors in a risky situation, especially if one has little time to decide what to do or when in panic [27]. People tend to have limited, or even incorrect, knowledge about fire situations, which makes them insufficiently prepared for a fire [27]. It is evident that to able to select the optimal risk-mitigating actions, one must have knowledge of the risk and be aware of possible actions and their implications (e.g. water cannot be used to extinguish a grease fire but a fire blanket can) [38]. Thus, in order to stimulate people to make the right preventive choices, increasing people’s knowledge level of fires and fire prevention might be important. An effective way to sustainably increase knowledge is by an emotional experience [3941]. Particularly negative experiences enhance the memory of that experience and its details [42]. The literature strongly suggests that serious games can be an effective way for learning [4345] because games are a good way to attach emotions to problem solving and with that, enhance learning [46]. Several empirical studies have indeed shown positive effects of serious games on knowledge. Kato, Cole, Bradlyn, and Pollock [47] have shown a positive effect of a 3-month serious game on knowledge about cancer treatment, compared to a control group who received a commercial game video game (i.e. Indiana Jones). One-time interventions have also proven to work: Wong et al. [48] have shown that a serious game is better in transferring factual knowledge compared to traditional textbooks, both directly after the intervention as one week later. A study of Chittaro [19] has also shown positive effects of a serious aircraft evacuation game in an IVE on knowledge retention. One week after the intervention the players of the game had a significantly higher score on the knowledge items than the group who had read a safety card, even though immediately after the intervention there was no difference between the groups. Furthermore, Chittaro [32] showed that an interactive mobile VR application teaching how to don a life preserver had positive effects on donning a life preserver in real life, comparing a traditional safety card.

Consequently, we expect that a serious game in an IVE, one that triggers emotions and increased attention by a virtual fire, can positively influence people’s knowledge level about fire prevention. We also expect that an increase in knowledge positively influences prevention behavior.

  • H1a: The knowledge level of participants in the IVE condition is higher than the knowledge level of participants in the INFO condition.

  • H1b: Knowledge positively influences prevention behavior.

Safety Locus of Control (SLOC)

Locus of control is an important psychological determinant that can influence our attitudes and behaviors [49] and for that reason it is often studied in situations where the goal is to change people’s behavior. Locus of control is our perception of where control lies, and most researchers distinguish between an internal and an external orientation [31,50,51]. Someone with an internal orientation is more likely to consider outcomes as a consequence of their own behavior, as opposed to someone with an external orientation, who often sees outcomes as something beyond their own control. In the literature there is no consensus about the dimensionality of the locus of control construct [50,52]. Although the original scale of Rotter [53] presents a one-dimensional construct (where internal and external are two ends of a continuum), later studies reveal two (with internal and external being two separate constructs) or even multiple dimensions. Most studies that we came across find empirically that locus of control is a two-dimensional construct, rather than the more intuitive one-dimensional construct [31,50,51]. The underlying argument is that people for example can attribute their health both to internal beliefs (for example related to their own smoking behavior) as well as to external beliefs (for example to chance events). An internal locus of control is associated with safer attitudes and behaviors: the more someone feels that he or she is responsible for how matters are progressing, the more likely that someone will take action. Inversely, an external locus of control is associated with a lack of caution and prevention. Studies have indeed shown that, for instance, people with an internal locus of control compared to people with an external locus of control are more likely to wear seat belts [28] and have less accidents at work [29,30]. Furthermore, studies showed that internal control is negatively related to fatal car accidents while external locus of control is positively related to fatal car accidents [31]. Huang and Ford [36] have shown in addition that locus of control is not a fixed human characteristic but can be influenced by training, as they have demonstrated with respect to safe driving behaviors. Murray, Fox, and Pettifer [54] have shown that a higher sense of realism in a virtual environment can likewise lead to an increased perception of locus of control. Also Ahn, Bailenson, and Park [26] managed to influence environmental locus of control with an IVE in which deforestation was at stake. Furthermore, in the domain of air safety Chittaro [17] showed that a videogame that focused on the brace position during a plane crash resulted in a significant increase in internal orientation and a significant decrease in external orientation, which together indicated that participants felt more in control over the outcomes of an emergency landing than before playing the game.

As far as we know, locus of control has not yet been analyzed in the fire safety domain, but it makes sense that it is relevant in that domain as well. We follow the line of reasoning that safety locus of control (SLOC) is a two-dimensional construct, existing of an internal orientation (ISLOC) and an external orientation (ESLOC). We expect ISLOC and ESLOC to be affected more for IVE than for the INFO condition and expect changes in ISLOC and ESLOC to positively influence behavior.

  • H2a: Perceived ISLOC is higher for participants in the IVE condition than for participants in the INFO condition.

  • H2b: Perceived ESLOC is lower for participants in the IVE condition than for participants in the INFO condition.

  • H2c: Perceived ISLOC positively influences prevention behavior.

  • H2d: Perceived ESLOC positively influences prevention behavior.

Vulnerability

An important determinant to take preventive measures according to the HBM and PMT, is the likelihood that an event will occur, referred to as the perceived vulnerability (or: susceptibility) [20,24]. A common barrier to take preventive measures is that people tend to think that bad things will not happen to them, a phenomenon known as ‘unrealistic optimism’ [7]. A way to increase people’s perceived vulnerability is through personal experience [7]. A study of Zaalberg et al. [6] has shown that people who have experienced a flood before, perceive themselves as more vulnerable to a future flood. This higher perceived vulnerability, together with a higher perceived effectiveness for adaptive actions, made that flood-victims had more intentions to take adaptive actions (e.g. tie up or remove curtains to prevent them from getting wet) than non-victims. However, no effect was found with respect to preventive measures (e.g. sandbags in front of the house). This might be explained by the fact that victims perceived the adaptive actions to be more effective than non-victims, while the effectiveness for preventive actions was found more effective by non-victims. That is, increased vulnerability leads to more preventive or adaptive measures, provided that the measures are considered effective enough.

Since we want to prevent people to become victims in the first place, we could increase the perceived vulnerability through the simulation of a risk. This has also been done by Schwebel et al. [55] who studied the effects of the virtual simulation of crossing a street while texting, and Chittaro [1518] who measured the effects of the simulation of an aircraft emergency and a terror attack. While the IVE’s concerning street crossing and a terror attack showed significant positive effects on perceived vulnerability [15,55], the IVE’s that simulated an aircraft emergency [1618] showed no significant effects. While the perceived probability of an aircraft evacuation might not increase through a virtual experience given that most people fly on airplanes relatively rarely, affecting people’s perceived vulnerability to a fire is more likely, as most people cook several times a week, or use lighters, candles etc. much more regularly. We expect that the virtual experience of a fire increases people’s perceived vulnerability to a fire, which in turn will result in more preventive behaviors.

  • H3a: The perceived vulnerability of participants in the IVE condition is higher than in the INFO condition.

  • H3b: Perceived vulnerability positively influences prevention behavior.

Severity

Another important determinant for taking preventive measures according to the HBM and PMT, is the perceived severity of the event and its consequences [20,24]. The direction of this relationship depends on the expected severity beforehand, since some experiences appear to be milder than expected [7]. With respect to flood experiences, research has shown that the experience of a flood did increase the perceived severity of a future flood, compared to people who did not have this experience [6]. Chittaro [17,18] managed to increase perceived severity in an aircraft evacuation game using rich and vivid feedback to induce fear, although the effect was small. In the domain of fire safety, increasing perceived severity is even more relevant (an airplane crash is already considered quite severe) since with fires in buildings people seem to downplay the severity and as a consequence do not respond quickly enough, which decreases their chances of survival [27]. Chittaro and Zangrando [14] argue that people’s fire evacuation behavior can be influenced by increasing anxiety and risk perception. Their study showed that a game about the dangerous effects of smoke had more impact on anxiety and attitude when emotional intensity was high (as it might be in an IVE). We therefore expect that a fire in an IVE can similarly increase perceived severity, which itself is considered a relevant psychological determinant for prevention behavior.

  • H4a: The perceived severity of a fire by participants in the IVE condition is higher than in the INFO condition.

  • H4b: Perceived severity positively influences prevention behavior.

Self-efficacy

Another important determinant to take preventive measures that is related to the person (instead of related to the risk) according to the extended versions of the PMT and the HBM is self-efficacy [23,25], a construct that is related to locus of control. “Perceived self-efficacy is the belief in one’s competence to tackle difficult or novel tasks and to cope with adversity in specific demanding situations” ([56], p. 81]. In this sense, self-efficacy is different from locus of control: it does not consider the extent to which a person feels that the situation depends on his or her own behavior, but instead how well a person perceives he or she would perform in the part that does depend on his or her behavior. According to Bandura [57] optimistic efficacy beliefs lead to better performance outcomes, those beliefs can be acquired by gaining experience, and this experience can be gained through an IVE [8]. For example, in the healthcare domain, serious games for children, compared to a control group, have shown positive effects on self-efficacy in taking care of a chronical condition [58,59]. Also in the safety domain similar effects are found: a serious aircraft evacuation game has shown positive effects on self-efficacy in safely evacuating an aircraft [16,18] and a mobile VR application concerning life preserver donning increased people’s self-efficacy [32]. Sometimes, instead of pursuing an increase in people’s self-efficacy level, it is desirable to decrease self-efficacy. For instance, in driving studies people tend to overestimate their driving ability [60] and in this case a higher perceived self-efficacy is related to more unsafe driving behaviors and accidents [6163]. This is comparable to the fire prevention domain where people generally tend to overestimate their ability to evacuate, because people downplay the severity of a fire [27,64]. In general, people start moving too late and too slowly in the case of a fire, and often even move through smoke while this should be avoided as this slows down their speed and is dangerous for their health. The suggested negative relationship between perceived severity and self-efficacy has been empirically shown in the health domain [65,66]. Since we expect that people underestimate the consequences of a grease fire, and overestimate their own ability to act properly, we expect that the IVE can decrease perceived self-efficacy with respect to a grease fire, both directly as well as through an increase in perceived severity and that a lower self-efficacy will motivate people to take more fire prevention measures.

  • H5a: The perceived self-efficacy of participants in the IVE condition is lower than in the INFO condition.

  • H5b: Perceived severity negatively influences self-efficacy.

  • H5c: Perceived self-efficacy negatively influences prevention behavior.

Benefits and barriers of behavior

Other determinants that drive prevention behavior according to the HBM and PMT are the benefits of the behavior (more specifically: its effectivity) and the barriers of the behavior (e.g. financial costs, effort). Because in our experiment a fire blanket is promoted as an effective measure to extinguish a grease fire in both conditions, we do not expect ‘effectivity’ to be influenced more in the IVE condition and therefore do not include effectivity in our study. Because the IVE and INFO condition did not stress the costs and effort involved concerning a fire blanket, we also do not expect costs and effort to play an important role in our study, and therefore do not include these determinants in our study.

Prevention behavior

The ultimate goal of the simulation of a fire in an IVE is to increase people’s fire prevention behavior. Most studies have only considered the effect of IVE on psychological determinants and sometimes on intentions but did not measure subsequent behavior. This implies we do not have much empirical support for an effect of IVE on prevention behavior. However, if our earlier hypotheses about the effects of IVE on psychological determinants are valid and changes in psychological determinants lead to changes in behavior, then the logical consequence is that IVE influences behavior.

  • H6: Because of the changes in psychological determinants caused by the IVE condition, we expect that participants in the IVE condition display more prevention behaviors than participants in the INFO condition.

Retention of the psychological determinants

Most IVE-related effect studies only consider effects immediately following the intervention. However, in some cases longer term effects on knowledge have been studied and have led to positive results for IVE. In Chittaro [19] we find that while there no was no immediate effect of a serious aircraft evacuation game on knowledge, the players of the game had a significantly higher score on knowledge than the control group one week later. Furthermore, Wong et al. [48] have shown that a serious game was better in transferring factual knowledge than a traditional textbook, and this effect was still demonstrable one week later. The sometimes implicit and largely intuitive argument for this effect is that because people have an increased sense of presence in an IVE (compared to people who read an information sheet), their experiences make a more lasting impression, which improves retention. Along these lines, we expect a higher retention rate (or slower decrease of retention) of the effects of the IVE on the psychological determinants.

Method

Design

We used a 2 factor (IVE versus INFO) between-subjects design, with 3 time points at which we asked the participant to fill out a questionnaire. The first questionnaire was right before the intervention, the second questionnaire right after the intervention and the third questionnaire four weeks after the intervention. Our main intervention has two levels: the IVE fire game (IVE) and a fire prevention information sheet (INFO) as a control.

Participants

Participants were recruited from the Dutch commercial “CG Selections” consumer panel, a panel consisting of 80,000 consumers. We determined that we would need 2 x 119 participants to have 90% power of showing an expected difference in prevention behavior of about twenty percentage points (assuming 20% for the INFO group vs 40% for the IVE group; using an alpha-level of 5%).The prevention behavior that we targeted was whether a participant would invest part of his or her show-up fee in a fire blanket and whether a participant would take home flyers related to fire safety. We recruited participants based on the following criteria: between 18 and 70 years old, not living in a student home or living with their parents (due to prevention responsibility), who did not own a fire blanket and who did not suffer from motion sickness (to prevent them from getting nauseous during the IVE). CG Selections approached potential participants based on the age and location information in their database with the opportunity to participate in a study about “property damage” for an incentive of €25. We did not mention the topic “fire prevention” to prevent any bias (e.g. attract people that have a special interest in fire; people becoming extra aware of their own fire prevention). To determine whether people fitted the required criteria to engage in the study they first had to fill out a survey with questions concerning their living situation, motion sickness, and whether they owned a fire blanket. Because we did not want to prime people with the fire blanket, we confronted people with a list of possible home appliances and asked which products they owned.

After enrolling, participants were randomly assigned to the IVE group or the INFO group. Each day was either an IVE day or an INFO day, and participants were assigned to a condition based on availability. In total 297 participants started the experiment, of which 49 were omitted from the analysis because in the first questionnaire (right before the intervention) they reported to own a fire blanket, three were excluded because they failed to fill out the second questionnaire and three others were excluded because they did not complete the IVE due to nausea. In total 242 participants remained in the dataset for analysis.

Procedures

The experiment was approved by the ethical committee of the Department Human Technology Interaction of the Technical University of Eindhoven. The experiment ran in the lab for nine days. In total nine persons assisted the first author in running the experiment. When entering the lab, first the participant was asked to read and sign a written informed consent. Second, we asked participants to fill out the first online questionnaire. Then, dependent on the assigned condition, the participant was asked to play the IVE game or read the INFO sheet. Participants in INFO condition received fire prevention information on a single page A4 (S1 File) with basically the same content as the experiences in the IVE contained (S2 File). The intervention lasted about 8 minutes for the IVE group and 2 minutes for the INFO group. To evaluate the IVE game, participants were asked afterwards whether they became nauseous or dizzy, how realistic they considered the fire experience, and how severe they considered this virtual experience. Then, participants in both groups were asked to fill out a second online questionnaire. Then, we offered the participant a choice of receiving their promised €25 or receiving €12.50 and a certified fire blanket with a value of €20. The money would be transferred after they filled out the third questionnaire (four weeks later), the fire blanket could be taken home immediately. In addition, to measure the interest for fire prevention information, there were two types of fire prevention related flyers on the table, which participants could take home. To avoid social desirability bias the availability of these flyers was not specifically mentioned by the experimenter. Four weeks after the participant was exposed to the intervention, the participant received the third online questionnaire.

The IVE condition and participant’s actions

In the IVE condition, participants were told that they were going to do a fire drill in a virtual environment. The participant received a game controller, a headphone, and a head-mounted-display. It was explicitly stated that if the participant became nauseous or otherwise uncomfortable, he or she could stop at any time. First, the simulation showed a practice scene so that the participant could get acquainted with the virtual environment. During the practice scene it was explained to the participant how to move in the environment (walk through the house, go up- or downstairs) and how to select an action (use water or fire blanket, open doors, pick-up toddler). The participant can move (walk, turn) in the IVE by using the joystick of the controller and choose different actions by focusing on the objects with his eyes. If the participant focusses on an object (e.g. the fire blanket), a blue lightbulb appears that is counting down from 5 seconds to 1 second, after which the action is performed (e.g. use fire blanket to extinguish fire). When the participant indicated to be ready for it, the actual game was started. See Fig 2 for screenshots of the IVE fire game.

Fig 2. Screenshots of the IVE fire game.

Fig 2

(A) View of the grease fire. (B) View of the bucket of water and the fire blanket that can be selected to extinguish the fire. (C) View of the flash fire. (D) View of the toddler that is waiting to be rescued.

At the start of the game the participant (in the IVE) is situated on the couch in the living room, watching television. Suddenly the smoke detector goes off, and smoke is entering the living room. The participant can 1.) go towards the source of the fire (the kitchen) and extinguish the fire with the fire blanket. The fire will be extinguished and the game ends. Or 2.) go towards the source of the fire (the kitchen) and extinguish the fire with water. In this case a flash fire occurs and the participant has to escape the house to survive (using the fire blanket at this point will not help any more). A final option was 3.) to escape from the home immediately through the front door. The participant could however also not perform action 1, 2 or 3, but stay in the house. Thirty seconds after the smoke detector goes off, the fire gets bigger as the flames now hit the kitchen hood and cabinets. At this point, if the participant is still in the house, he will hear a toddler crying upstairs, and he could choose to go upstairs and rescue the child. After the participant had either extinguished the fire, escaped the home or 105 seconds had passed, the game ended. Dependent on the decision the participant made, textual feedback and information about fire prevention was presented. For instance, if the participant had not been able to extinguish the fire and was still in the house, a text would explain that in case of a fire in all likelihood three minutes is the maximum one has to escape, and that proper use of the fire blanket can prevent a lot of harm. See S2 File for the possible scenarios and the feedback given after each scenario.

After playing the game for the first time, participants were asked to play the game once again. If they had extinguished the fire with the fire blanket the first time, they were now suggested to play the game without using the fire blanket (in which case the optimal strategy would to try to escape). If they did not extinguish the fire with the fire blanket the first time, they were asked to try to extinguish the fire with the fire blanket. This way, all participants experienced both escaping and extinguishing a fire.

Hardware

We used the Oculus Rift DK2 Head-Mounted Display (HMD) with a resolution of 960 x 1080 per eye and a 100 degrees diagonal field of view. Audio was played through Sennheiser HD 265 linear headphones. For the study, the software was implemented on one desktop machine and one laptop. Given our setup, the specifications of the hardware needed to be substantial, given current day hardware. The desktop machine was a Mac Pro 3.7GHz, with 16GB of RAM, a 256GB Flash Drive and 2x AMD Fire Pro D300 graphics card with 2GB each. The laptop was a MacBook Pro 2.4Ghz, with 8 Gb of RAM, 1 TB HDD, and a GeForce video card with 512MB.

Measurements

For this study we used measurements at three time points. The first measurement was a questionnaire right before the intervention that included background questions and current prevention behaviors that are beyond the score of this paper. Background questions included family composition, type of home, ownership of home, year of construction, frequency of cooking, frequency of playing computer games and experience with home fire(s). The second measurement took place right after the intervention and was set up for the measurement of the psychological determinants and the target prevention behaviors. The hypothesized model (Fig 1) is based on differences between the IVE and the INFO group at the time of the second measurement. The third measurement, four weeks after the intervention, was set up to be able to analyze the development of the psychological determinants over time. In this measurement, the determinants knowledge, vulnerability and severity were included, and we also included the same prevention behaviors as in the first measurement. We excluded self-efficacy, ISLOC and ESLOC from the third measurement because of lengthiness, which could negatively affect responses to the (at-home) questionnaire and reliability of the answers. The prevention behaviors that were measured in the first and third measurement are not analyzed in this paper.

Knowledge

Knowledge concerning fire prevention was measured with nine self-constructed items that were related to the information provided in the IVE fire game and the INFO sheet. Example items are: What happens when you throw water on a grease fire? and On average, how many minutes does a person have to safely leave the home in case of a fire? See S3 File for an overview of all items. Correct answers were coded 1 and incorrect answers were coded 0. The overall knowledge score was the percentage of correctly answered items.

Internal and external safety locus of control

We designed the locus of control scale specific for fire prevention, since a domain-specific locus of control scale is a better predictor for domain specific behavior than the general locus of control scale [67]. In line with Chittaro [17] locus of control was measured with 12 items, which were adopted from Hunter [51], and were adapted to fit the topic of fire. There were 6 items internally oriented (ISLOC) and 6 items externally oriented (ESLOC). What complicates matters somewhat is that locus of control can refer to two separate matters in the event of a fire: whether someone feels that he or she can influence the probability of a fire, and whether one can influence the consequences of a fire. We formulated items for both matters. An example of an internally oriented item that was focused on preventing a fire was: If you are careful, you can prevent fire in your home. An example of an externally oriented item that was focused on reducing the consequences of a fire was: If there is a fire in your home, there is usually nothing you can do. Ratings were given on a five-point Likert scale (1 = fully disagree; 5 = fully agree).

Vulnerability

Vulnerability was measured with 3 items. In line with Chittaro [16,17] we adopted the items from de Hoog, Stroebe, and de Wit [68], and transferred the items to the subject of a grease fire. Ratings were given on a five-point Likert scale (1 = very low; 5 = very high). An example item was: How high do you perceive the risk of a grease fire in your home to be?

Severity

Severity was measured with 3 items. In line with [16,17] we again adopted the items from de Hoog et al. [68] and transferred the items to the subject of a grease fire. Ratings were given on a 5-point Likert scale (e.g. 1 = not severe at all; 5 = very severe). An example item was: How severe do you perceive the consequences of a grease fire to be? We changed one item of original scale, namely concerning the seriousness of the grease fire, since in Dutch this would also be translated as “severe”. We changed the item to “panic” since this also reflects the severity of the grease fire.

Self-efficacy

Self-efficacy was measured with 10 items. The items were adopted from the General Self-Efficacy Scale (GSE) [69] and from the scale that Chittaro [16] used to measure self-efficacy with respect to aircraft evacuation (which was also based on the GSE). All 10 items were transferred to the domain of fire prevention behavior, following [70] who argues self-efficacy should be tailored to the specific domain of functioning to have the best explanatory and predictive value. Ratings were given on a five-point Likert scale (1 = fully disagree; 5 = fully agree). An example items was: I am confident that I can extinguish a grease fire.

Specific prevention behavior

In both conditions of the experiment a fire blanket is the advised behavior to extinguish a grease fire. The primary measurement for the effect of IVE on prevention behavior will be the participants’ purchase of a fire blanket, and the secondary measurement is the participants’ interest in fire prevention information (which is a more ‘soft’ measurement of behavior). We measured the purchase of a fire blanket by offering the participant a choice between the promised incentive of €25, or €12.50 and a fire blanket for participation. Concerning fire prevention information, there were two types of prevention related flyers on the table that participants could take home. We registered (unobtrusively) whether a participant chose the fire blanket (0 = no fire blanket; 1 = fire blanket) and whether the participant took one (or more) flyers (= no flyers; 1 = flyers).

Additional measurements

For the experiment we only recruited people that reported not to own a fire blanket. To verify, we asked again if people had a fire blanket in the first questionnaire. If so, they were excluded from the analysis. Furthermore, we included several variables that we used for evaluation purposes and robustness checks: we verbally asked the IVE group right after the IVE to what extent they became nauseous or dizzy, how realistic they considered the fire experience, and how severe they considered the virtual experience.

Results

First we briefly describe our preparation of the data and then show some descriptive statistics. To answer our research question with respect to the impact of IVE on the psychological determinants and the actual prevention behavior, we use Structural Equation Modelling (SEM). Finally, we analyze the development of the psychological determinants between the second and the third measurement.

Data preparation

There were no missing values in the first and the second questionnaire, except as a consequence of conditional items. However, 10.6% of the participants (26 out of 242) did not fill out the third (at home) questionnaire. We checked the data for multivariate outliers with the BACON algorithm using Stata 14 [71] and did not find any. We checked the data for normality with a skewness and kurtosis test [72] and the Shapiro-Wilk test [73] and found that not all variables were normally distributed so we used a robust estimator for non-normally distributed data. For SEM it is important that items or latent constructs do not correlate too much with each other [74] and there were no bivariate correlations larger than .85. Detecting multicollinearity among multiple variables was done by considering the variance inflation factors (VIF) after a logistic regression with the choice of a fire blanket as the target variable and all scale constructs included. This assumption was not violated: all variables have a VIF <10 and together have a mean VIF of 1.20 [73].

Descriptive statistics

Of the remaining 242 participants in our analysis, 124 were assigned to the IVE condition, and 118 to the INFO condition. Of these 242, 146 females (60.3%) and 96 males (39.7%). Mean age equaled 42.5 (SD = 10.11). A chi-square test and a two-sample Wilcoxon rank-sum test (age was non-normally distributed) showed that there was no significant difference between the IVE and INFO condition in terms of gender and age respectively (χ2 (1, N = 242) = .703, p = .402; z = .164, p = .870).

After the IVE game, 21.8% indicated that they had felt nauseous or dizzy during the experience, 32.3% indicated that they felt a little nauseous or dizzy, and 46.0% did not feel nauseous or dizzy. Participants rated the realism and severity of the IVE experience on a five point Likert scale (where a five stands for very realistic / severe). The mean scores were 3.82 (SD = .97) for realism and 3.43 (SD = 1.11) for severity. In the IVE condition, 48.4% of the participants chose the fire blanket and €12.50 and in the INFO condition this was 39.8% (χ2 (1, N = 242) = 1.795, p = .180). In the IVE condition, 23.1% of the participants took the flyers home and in the INFO condition this was 12.8% (χ2 (1, N = 242) = 4.179, p = .041). So there is some evidence for a relation between the IVE manipulation and prevention behavior, but the effect is smaller than we assumed (a twenty percent point difference) when calculating our sample size. The specific actions that people took in the IVE fire game can be found in S1 Table.

Statistical analyses

We tested the model in Fig 1 through SEM with a robust estimator for non-normally distributed data. We used SEM because we wanted to determine the relations between the observed and latent constructs in a single estimation. An alternative possibility to test the hypothesized model is to analyze the model with SEM in three steps, namely first test the total effect, then the effects of IVE on the psychological determinants followed by the effects of the determinants on prevention behavior, in line with a standard mediation analysis. Or, we could separately show the effect of VR on the psychological determinants, and only then the (separate) effects of the psychological determinants on behavior. We found that these separate analyses show substantially the same results: all hypotheses of interest remain statistically significant and of similar size. For parsimonious reasons, we present the SEM model that tests all the relationships in a single estimation.

Some further beneficial features of SEM are that it is possible to use both observed and unobserved (i.e. latent) variables, and that the measurement error of observed and unobserved variables are taken into account simultaneously. Different compared to most experimental IVE studies is that with SEM the model as a whole is being tested instead of a set of individual hypotheses, including the role the psychological determinants might play in the effect of IVE on prevention behavior. The SEM model was analyzed in Mplus version 8 [75]. The estimation procedure that we used was the default weighted least squares with means and variance adjusted (WLSMV, which implies using probit as the underlying analytical model) since this is considered the best estimator for categorical data [74,76]. Whether the model fits the data can be determined through the model fit statistics and the individual parameter estimates. As advised in Hair, Black, Babin, and Anderson [77] we report the following model fit statistics: model chi-square, root mean square error of approximation (RMSEA) and its 90% confidence interval [78], the comparative fit index (CFI) [79] and the Tucker Lewis Index (TLI) [80]. The higher the model chi-square the worse the fit, and the associated level of significance must be non-significant [77]. For the RMSEA Hu and Bentler [79] advise a value smaller than .06 for a good fit, with the upper bound of its 90% CI falling below 0.10 [81]. For the CFI and the TLI, Hu and Bentler [79] advise a value larger than .95.

Confirmatory factor analysis

The latent constructs were analyzed by CFA in a single estimation, so that correlations between (items of) latent constructs could be taken into account [81,82]. When all items of the latent factors (vulnerability, severity, self-efficacy, ISLOC and ESOC) were used, model fit statistics were poor: χ2 (340) = 1080.666, p = < .001, CFI = .875, TLI = .861, RMSEA = .095, 90% CI [.089 - .101]. In order to establish construct validity we inspected the significance levels of the items, the direction of the estimates, the item estimates and R-squares (83). Convergent validity and a more precise measurement is established by removing items based on the AVE, which ideally has to be larger than .5 [77]. Indeed, inspecting the CFA we find that many items have low R-squares and that the AVE of the constructs was low (below .5 for some). We increased the AVE for self-efficacy from .560 to .682 by removing five items, for ISLOC we increased the AVE from .364 to .413 by removing three items, and for ESLOC we increased the AVE from .343 to .549 by removing four items. Since ESLOC only had two items left, and a minimum of three indicators per latent variable is needed for model identification [82] the correlation between ISLOC and ESLOC was very high (-.792, suggestion discriminant validity is compromised as the correlation is higher than the square root of the AVE of ISLOC itself), and the items seem to fit well on one scale, we decided to use a one factor SLOC instead. When performing CFA with all latent constructs and the latent factor SLOC, the same results apply concerning item deletion, as with CFA with all latent constructs and the latent factors ISLOC and ESLOC. Internally orientated items (I) contribute positively to SLOC, while externally orientated items (E) contribute negatively to SLOC. Although in literature two factors are more common, LOC was originally developed as a one-dimensional construct and is still used as such by various studies [26,53,83]. Table 1 shows the final AVE and Cronbach’s alpha values of the latent factors, and the factor loadings per item. The model fit then improved to χ2 (98) = 478.448, p = < .001, CFI = .914, TLI = .895, RMSEA = .127, 90% CI [.115 - .138]. The factor loadings and the Cronbach’s alpha values of the original scales can be found in S4 File, Table 1. In addition, we performed CFA on the individual factors and similar results apply compared to testing the measurement models all together, and the same items should be removed to improve AVE. See Tables A-E in S4 File for the model fit statistics and Tables F and G in S4 File for the R-squared estimates and AVE’s.

Table 1. AVE, Cronbach’s alpha and standardized factor loadings (stdYX) for the latent factors, after removing poorly fitting items.

Items without a factor loading where excluded from analysis.

Latent factora Itemb Factor loadingc (stdYX)
Vulnerability How high do you think the risk of a grease fire in your home is? .570
α = .745 How high do you perceive the chance that a grease fire will pass .660
AVE = .584 over to the exhaust hood and the kitchen cabinets?
How high do you perceive the chance that you should escape your home because of a grease fire? 1.002
Severity How dangerous do you think a grease fire is? .829
α = .752 How severe do you perceive the consequences of a grease fire? .851
AVE = .705 How much panic do you think there will be in case of a grease fire? .840
Self-efficacy I am confident that I can extinguish a grease fire. .834
α = .834 I am confident I will remain calm in case of a grease fire. .848
AVE = .682 I am confident I will remain calm in case of a grease fire, even if it will pass over to the exhaust hood and the kitchen cabinets.
When a grease fire exists I am afraid I will panic. d .782
I know what to do in case of a grease fire. .772
I am capable of acting correctly in case of a grease fire. .889
I am convinced of my capability to put my family/ myself into safety in case of a fire.
I am convinced of my capability to quickly leave my home in case of a fire.
I am convinced of my capability to quickly leave my home in case of a fire, even if escape routes are blocked.
I am convinced of my capability to quickly leave my home in case of a fire, even if there is a lot of smoke.
Safety Locus of If you are careful, you can prevent a fire in your home yourself. (I)c
Control (SLOC) If a fire breaks out in your home, there is usually nothing that you
α = .710 can do. (E)e
AVE = .528 A home fire is usually caused by a short-circuit/ overheating of
electrical appliances. (E)
Whether people can escape in time in case of a fire, is a matter of luck or bad luck, not of preparation. (E)
People can ensure that a small fire does not expand. (I) .653
Most home fires are caused by chance events that cannot be influenced. (E)
Preparing yourself for a fire, enlarges your survival possibilities in case of a fire. (I) .661
Whether you succeed in extinguishing a grease fire, is a matter of luck or bad luck, not a matter of preparation. (E) -.777
People should be rewarded by their insurance company if they take preventive measures to prevent or control a fire. (I)
By taking preventive measures you can make sure that you can extinguish a fire on time. (I) .733
Home fires are usually caused by the people themselves. (I)
It has no use to prepare yourself for a fire in your home. (E) -.796

a Latent factor originates from the second measurement.

b Items are translated from Dutch.

c Factor loadings are only present for items that were not removed from the scales.

d This item was reversed coded.

e The (I) refers to an internally orientated item. The (E) refers to an externally oriented item.

Structural Equation Models

We now test the model as depicted in Fig 1. Table 2 shows the results of fitting six different models, each slight variations of the base model from Fig 1. The models differ in terms of the changes that were made to improve model fit, but do not change in terms of the significance and size of the effects for our underlying hypotheses. That is, the general conclusions with respect to our hypotheses do not depend on, which model is used, underscoring the robustness of our findings.

Table 2. Model fit statistics, standardized regression weights [stdYX] and standard errors [S.E.] for model 1–6.

Model fit statistics
Goodness of fit Target values Model 1 Model 2 Model 3 Model 4 Model 5 Model 6
χ2 590.867 408.292 150.458 390.103 148.994 15.753
df 78 76 73 81 78 12
p >.05 < .001 < .001 < .001 < .001 < .001 .203
RMSEA < .06 .165 .134 .066 .126 .061 .036
90% CI < .10 .153-.177 .122-.147 .051-.081 .113-.138 .046-.076 .000-.079
CFI >.95 .854 .905 .978 .912 .980 .994
TLI >.95 .803 .869 .968 .886 .973 .990
Relationships
Independent variabele Dependent variabele StdYX [S.E.] StdYX [S.E.] StdYX [S.E.] StdYX [S.E.] StdYX [S.E.] StdYX [S.E.]
IVE knowledge -.333[.060] -.333 ***[.060] -.333***[.060] -.333***[.060] -0.333***[.060] -.333***[.060]
IVE vulnerability .229**[.069] .231**[.067] .231 **[.068] .240***[.067] .240***[.067] .239**[.069]
IVE severity .169*[.069] .169*[.069] .158*[.071] .169*[.069] .158*[.071]
IVE self-efficacy -.231***[.064] -.230***[.064] -.276***[.066] -.230***[.064] -.276***[.066]
severity self-efficacy -.387***[.054] -.388***[.054] -.189 **[.068] -.388***[.054] -.189**[.068]
knowledge fire-blanket .225**[.086] .225**[.086] .225**[.086] .225**[.086] .225**[.086] .225**[.086]
vulnerability fire-blanket .124[.123] .091[.102] .078[.103]
severity fire-blanket .136[.118] .102[.101] .096[.095]
self-efficacy fire-blanket .049[.118] .074[.107] .033[.102]
knowledge flyers .079[.659] .225[.086] .010[.084]
vulnerability flyers .321*[.138] .246*[.110] .245*[.111] .251*[.103] .252*[.103] 0.264**[.102]
severity flyers -.024[.121] -.057[.099] -.052[.094]
self-efficacy flyers -.143[.144] -.032[.128] -.030[.128]
IVE fire-blanket .154[.092] .167[.088] .160[.088] .183* [.084] .183*[.084] .183*[.084]
IVE flyers .228[.220] .143[.100] .142[.101]
Independent variable Independent variable
fire blanket flyers .488 ***[.107] .490 **[.106] .490**[.101] .505***[.108] .505***[.108] .507***[.108]
severity vulnerability .292***[.059] .287***[.060] .289***[.059] .285***[.061]
self-efficacy vulnerability -.324***[.058] -.395***.[55] -.325***[.058] -.395***[.055]

Model 1: as proposed in Fig 1 but without SLOC

Model 2: as model 1 and with correlations between the latent variables

Model 3: as model 2 and with three correlations on an item level

Model 4: as model 2 but without the non-significant relationships

Model 5: as model 3 but without the non-significant relationships

Model 6: as model 4 but without all relationships that did not have an [in]direct effect on the target variables

* = p < .05

** = p < .01

*** = p < .001

Model fit statistics of the initial model (Fig 1, not in Table 2) were poor: χ2 (155) = 807.382, p = < .001, CFI = .840, TLI = .804, RMSEA = .132, 90% CI [.123 - .141]. Since IVE did not significantly affect SLOC and SLOC did not relate to prevention behavior, we removed SLOC from further analyses. We then tested the model as proposed in Fig 1 again without SLOC (model 1). Model fit statistics improved but were still not adequate, as can be seen in Table 2. We further improved the model fit, as suggested by the mod-indices, by adding correlations between latent variables that we did not hypothesize a-priori (model 2) (severity ↔ vulnerability; self-efficacy ↔ vulnerability). There relations are not illogical since for example vulnerability and severity together are assumed to represent a larger construct referred to as “risk perception” [84] or “threat appraisal” in Protection Motivation Theory [24,25]. Also we added correlations between items (model 3) (self-efficacy ↔ sev15; self 6 ↔ self5; self6 ↔ self1), as these relations were suggested by a mod-indices analysis. We only incorporated this step to show how model fit could be improved, but will not elaborate on this step, as adding correlations at the item level is not very common. A logical next step, after model 2, is to inspect the significant relations and remove the non-significant ones (model 4). Some relations that were hypothesized a priori appeared to be non-significant and were removed, of which many represented the effect of psychological determinants on behavior (vulnerability → fire blanket; severity → fire blanket; self-efficacy → fire blanket; knowledge → flyers; severity → flyers; self-efficacy → flyers; IVE → flyers). Model fit indeed improved compared to model 2. For a visual representation of model 4 see Fig 3. Next, we removed the non-significant relations from model 3 and found that the model fit further increases but the relations themselves do not change (model 5). We then elaborated on model 4 and removed all constructs that have no direct or indirect effect on the target variables (fire blanket or flyers) (Model 6). Although this procedure results in a better model fit, the relationships between the observed and latent constructs remain the same, which underscores the robustness of our findings. Model fit statistics are then very good. For a visual representation of model 6, see Fig 4.

Fig 3. Structural Equation Model of model 4 with standardized regression weights [StdYX].

Fig 3

Squares represent observed constructs and oval shapes represent latent constructs. Significance levels: * = p < .05; ** = p < .01; *** = p < .001.

Fig 4. Structural Equation Model of model 6 with standardized regression weights [StdYX].

Fig 4

Squares represent observed constructs and oval shapes represent latent constructs. Significance levels: * = p < .05; ** = p < .01; *** = p < .001.

Robustness checks

We tested the hypothesized relationships also on variations of the original sample to check the robustness of the outcomes. We tested all the relationships on the sample minus the participants who stated they became nauseous during the IVE experience (n = 215), on the sample with only the IVE participants who stated to find the experience (very) realistic (n = 199), and on the sample with only the IVE participants who stated to find the experience (very) severe (n = 187). With the exception of some small deviations in model fit statistics and standardized regression weights, results showed that all estimated relationships remained stable across the different samples.

Hypotheses

Tests of the hypothesized relationships can be found in Table 2 and we will address them here in more detail based on Model 4 (Fig 3). Contrary to what was hypothesized (H1a), the IVE fire game results in a lower knowledge level than the INFO condition (-.333, p < .001). Thus, participants who were provided the text sheet knew more about fire prevention than participants who experienced the virtual fire and were provided with this same information in the IVE. As expected, an increase in knowledge implies a higher probability to choose the fire blanket (H1b) (.225, p = .009). H2a-H2d concerned ISLOC and ESLOC. Based on our results we integrated ISLOC and ESLOC into a single concept: SLOC, and as a consequence we could only test H2a and H2c. Further analyses showed that participants who experienced the virtual fire did not have a higher internal safety locus of control compared to participants who were provided with the text sheet. Also, a higher level of internal safety locus of control did not correlate with prevention behavior. Therefore, H2a and H2c are rejected. H3a is supported: the vulnerability of participants in the IVE condition is higher than the vulnerability of participants in the INFO condition (.240, p < .001). H3b can be partly confirmed, as an increase in vulnerability only influenced taking home flyers (.251, p = .015), but did not affected participants’ choice for the fire blanket. H4a is supported: the severity of a fire by participants in the IVE condition is higher than the severity by participants in the INFO condition (.169, p = .014). However, an increase in severity did not influence prevention behavior (H4b). As hypothesized in H5a, the self-efficacy of participants in the IVE condition is lower than the self-efficacy of participants in the INFO condition (-.230, p < .001). Also, a higher severity negatively influenced self-efficacy (H5b) (-.388, p < .001). However, a lower self-efficacy did not affect prevention behavior (H5c).

Strictly speaking, H6 is rejected, but some careful consideration is necessary here. To test H6 we now distinguish the total, direct, and indirect effects based on model 6 from Table 2 (Fig 4), using a percentile bootstrap estimation approach with 1000 samples [85]. All coefficients are presented as standardized regression weights (stdXY). The total effect of IVE on fire blanket consists of a significant direct effect of IVE on the fire blanket (b = .197, SE = .081, 95% CI [.065, .726], p = .016) and a significant indirect effect via knowledge (b = -.080, SE = .030, 95% CI [-.279, -.044], p = .007). The total effect, however, is non-significant (b = .116, SE = .078, 95% CI [-.082, .550], p = .136). Hence, the non-significant total effect of IVE on fire blanket appears to be a combination of two significant opposite effects (positive and negative). Additionally, the effect of IVE on flyers is the consequence of (only) an indirect effect via vulnerability. There is a positive and significant effect of IVE on vulnerability (see Model 6, Table 2) and a positive and significant effect of vulnerability on flyers (see Model 6, Table 2), but the total effect is only significant at the p = .052 level (b = .063, SE = .032, 95% CI [-.002, 254], p = .052). Moreover, H6 implies that any effect of IVE on prevention behavior will be mediated by the psychological determinants. This part of the hypothesis is clearly rejected for the fire blanket, as there remains a direct effect of IVE on choosing the fire blanket.

Retention after four weeks

We now consider the development of the scores on the psychological determinants, to test whether the IVE has a positive effect on retention. Results are analyzed with a multilevel regression analysis with robust errors, correcting for the non-normality of the residuals.

On average, the knowledge level had decreased after 4 weeks, as there was a significant difference on knowledge between the first measurement (M = .86, SD = .13) and the second measurement (M = .78, SD = .13; b = —.038, p < .001). This decrease in knowledge is much stronger for the INFO group (from M = .91, SD = .09 to M = .79, SD = .13) than for the IVE group (from M = .82, SD = .14 to M = .78, SD = .13), as reflected in a statistically significant interaction (b = -.096, p < .001). For a graphical illustration of these differences, see Fig 5.

Fig 5. Knowledge level of the IVE and the INFO condition in the first and second measurement.

Fig 5

The error bars represent the standard error (SE).

On average, the vulnerability level was lower after 4 weeks, as there was a significant difference with respect to vulnerability between the first measurement (M = 2.64, SD = .83) and the second measurement (M = 2.54, SD = .82; b = .327, p = .002). There was no significant interaction effect between IVE and vulnerability (b = .1636, p = .085) (see Fig 6).

Fig 6. Vulnerability level of the IVE and INFO condition in the first and second measurement.

Fig 6

The error bars represent the standard error (SE).

On average, the severity level was slightly lower after four weeks, but there was no significant difference on severity between the first measurement (M = 3.73, SD = .62) and the second measurement (M = 3.63, SD = .61; b = -.062, p = .218). There was no significant interaction effect between IVE and severity (b = -.063, p = .379) (see Fig 7).

Fig 7. Severity level of the IVE and INFO condition in the first and the second measurement.

Fig 7

The error bars represent the standard error (SE).

Whereas previous results in the literature have suggested increased retention in IVE, our results support this only to a limited extent. First, vulnerability decreases somewhat over the span of four weeks, but we do not find substantial differences between the IVE and the INFO condition. Severity does not significantly decreases over time. There is a difference over time in terms of knowledge retention. After four weeks, the higher knowledge that the participants had in the INFO condition has decreased to the level of the IVE condition, which itself hardly decreased over the four weeks. In this sense, there is some evidence for an IVE leading to better retention than INFO in terms of knowledge. However, in our case the IVE condition had a lower knowledge level to begin with.

Discussion

We considered the effects of an IVE on relevant and frequently used determinants for prevention behavior—knowledge, vulnerability, severity, self-efficacy, and locus of control—, and the extent to which these effects in turn affect subsequent prevention behavior. The main reason for our approach was that most IVE studies only consider the effect of the IVE on these psychological determinants [13,14,1619,34,86], but not its effects on actual prevention behavior, nor the mediating effects of the psychological determinants.

The IVE had effects on almost all of the psychological determinants. As expected, the vulnerability and severity levels were higher in the IVE condition compared to the INFO condition: people in the IVE group felt more vulnerable to fire and thought a grease fire was more severe. After four weeks these feelings were approximately equally present in both groups. Also, as expected and desired, the level of self-efficacy was lower in the IVE condition compared to the INFO condition. That is, people felt less confident that they would act properly in case of a grease fire. This is partly a direct effect, and partly an indirect effect through the increase of the perceived severity of a grease fire. An unexpected result was that people in the INFO condition scored higher on knowledge. This could possibly be explained by the fact that the information in the IVE was presented as text in the head-mounted display, which may not have been so easy to read and perhaps also not compatible enough with the rest of the virtual experience. Also, in the INFO condition people were really focused on the information, as their only task was to read the presented information, while in the IVE condition people were more focused on the fire and on how to act and interact. One could imagine that the effect of the IVE could be improved by changing the way in which information is put forward. For instance, improved software might make it easier to produce a static text that is easier to read than the one in our IVE was, or by getting the information across through sound. Also, instead of showing the information afterwards, it might work more effectively to integrate all knowledge components in the IVE game (as in: [19,32]). Even though this was done for some knowledge components (e.g. the text”water causes a flash fire” was shown after the participant used water to extinguish the grease fire), other knowledge components were only presented at the end. Although the INFO group scored higher on knowledge directly after the intervention, their knowledge level significantly decreased after 4 weeks and decreased more rapidly than in the IVE group. This suggests that knowledge presented in a traditional text format is not retained over a longer time span as well as information in an IVE, as in line with results of Chittaro and Buttussi [19]. Safety locus of control (SLOC) was not influenced by the IVE at all, even though effects of interventions (e.g. IVE, training) on locus of control have been found in other domains [17,36]. A possible reason for this might be that locus of control is hard to influence in the fire prevention domain, possibly because the score of SLOC was already quite high (the mean score of the SLOC items equaled 4.3 were all items were measured on a five point scale).

The IVE did affect the measured prevention behaviors, however it affected different behaviors in different ways. When only comparing the subsequent behavior and the original conditions, IVE did not have the desired positive effect on investing in a fire blanket, compared to a control condition in which the same information was delivered on paper. However, closer inspection of this result shows that this effect was a combination of a direct positive effect of IVE and an indirect negative effect via a decrease in knowledge. One could potentially argue that the negative effect of IVE on knowledge could have had consequences for the other psychological determinants, however we did not came across these relationships in literature, nor did the mod indices in our SEM models suggest this. IVE had a (small) effect on vulnerability and vulnerability had a (small) effect on taking home flyers. Taken together this lead to a (smaller) total effect of IVE on taking home flyers that was no longer significant in our sample. It is somewhat unexpected and noteworthy that the IVE affects these prevention behaviors in different ways. This highlights the importance of a better understanding of the causal mechanisms that might lead individuals to change their behavior. In this case we found that a higher knowledge level results in investing in a fire blanket, but not to more information seeking through taking home flyers. More knowledge about the risk, the available actions and its consequences results in the fire blanket as the perceived optimal choice. The explanation for no increase in flyers might be that because one already has an increased knowledge level, there is no interest in further information. On the other hand, we found that a higher vulnerability did lead to more information seeking. Perhaps the higher vulnerability is a sign that participants realized that they knew less about the risks that they thought they did, and therefore wanted to get more information, which is understandable, although we arrive at this only in hindsight. It remains to be seen whether this connection between vulnerability and information seeking holds across a broader set of domains. Nevertheless, it seems sensible for future research to categorize the prevention behaviors in terms of the kind of psychological determinant it is affected by, instead of assuming that all prevention behaviors are equally affected by the psychological determinants.

The IVE affected knowledge, vulnerability, severity, and self-efficacy but the latter two did not relate significantly to prevention behavior, despite the fact that previous literature has indicated these to be potentially important determinants of behavior. The effects on the prevention behaviors are only affected by two of the psychological determinants: knowledge and vulnerability, and both are determinants for different prevention behaviors. Moreover, the effect of IVE on investing in a fire blanket is not fully mediated by the decrease in knowledge, but is for the most part a direct effect that cannot be explained away by any of the psychological determinants that we considered. This raises the question which other factors there might be that could explain the direct relationship between IVE and prevention behavior that we did not take into account. One possibility would be to consider the other factors of the HBM and the PMT that we did not include here, namely whether the perceived benefits and barriers related to the prevention behavior in question might nevertheless have played a role [20,24]. One of these variables is for example the perceived effectiveness of the prevention behavior, an important determinant, together with vulnerability, for taking adaptive actions to minimize the consequences of a flood [6]. Another question that arises because of the results of this study is how to interpret the effect of IVE on psychological determinants in other studies. Does an increase in for instance vulnerability or severity through IVE necessarily imply an actual behavioral change? Especially in fields in which the specific psychological determinants under study have not been validated with real behavioral outcomes (e.g. aircraft evacuation, flood experience, fire) our analyses suggests that one must be careful with the interpretation and generalization of such results.

Limitations

As the focus of this study was on fire and fire prevention, it remains to be seen whether results are generalizable to other areas of risk. For example, SLOC was not influenced by the IVE in our study, and did not affect the measured prevention behaviors, while studies in other domains did show these results.

Furthermore, we only measured two kinds of behavior: investing in a fire blanket and taking home flyers. As we have seen, direct and indirect effects may differ depending on the measured behavior and we do not know whether and how results would differ had we included additional prevention behaviors. For example, maybe an increase in the perceived severity would have influenced the probability of purchasing a fire extinguisher or would have led to paying more attention to the fire escape plan in the building.

In addition, one could argue about the appropriateness of the INFO condition as the control condition, as it differs on multiple aspects from the IVE condition (e.g. occupation time, no graphics, no gaming element). An alternative would have been to complement the current setup with a more comparable condition such as the 2D version of the same fire game, although this would probably lead to smaller sized effects. Our power analysis showed that we needed about 240 participants, so adding an additional condition would already dramatically increase the number of participants. Moreover, if the effect size would indeed be smaller given that the two conditions are more similar, this would imply a still larger necessary sample size.

We could also have measured the psychological determinants before the experimental manipulations, so that they could be used as control variables in our analyses, which might have increased precision and could have provided a baseline measurement against which to compare post-intervention measurements. The trade-off, however, is that we would be priming the participants in the direction of effects on these determinants by the intervention (possibly making it ‘easier’ to find differences). We therefore chose not to measure the determinants beforehand.

A further useful addition could have been to include the perceived effectiveness of the fire blanket as a measurement, since this is perceived to be an important determinant for performing prevention behaviors [7]. However, we did not included this question since we did not want to prime participants too much in the direction of buying the fire blanket.

Conclusion

Although the IVE influenced most of the psychological determinants, not all of these psychological determinants subsequently influenced the target prevention behaviors. The effect of the IVE on investing in the fire blanket was partly mediated by knowledge (and partly a direct effect), and the effect on taking home the prevention flyers was fully mediated by an increase in vulnerability. Given that we have included all psychological determinants that we found in the literature, it is surprising that the larger part of the effect of IVE does not seem to be correlated with these determinants. This suggests three issues that are noteworthy for IVE research in general. First, merely that only establishing effects of IVE on psychological determinants such as vulnerability and locus of control (as is usually the case in IVE studies) does not necessarily imply effects on prevention behaviors as not all psychological determinants lead to subsequent prevention behavior. Second, different prevention behaviors can be influenced by different psychological determinants: there is a real need to consider in more detail which determinants trigger which kinds of prevention behaviors. Finally, our study shows that not all effects of IVE on prevention behavior can be “explained away” by the psychological determinants that find their origins in the HBM and PMT and are typically measured in IVE research. Taken together, these two findings should be cause for some concern with regard to the often used setup in IVE research, where only the effect of IVE on psychological determinants is measured.

Supporting information

S1 File. Text for INFO group (translated from Dutch).

(DOCX)

S2 File. Text in IVE (translated from Dutch).

(DOCX)

S3 File. Knowledge scale (translated from Dutch).

(DOCX)

S4 File. Additional results of Confirmatory Factor Analysis.

(DOCX)

S1 Table. Actions people took in the IVE fire game, during the first and second game play.

(DOCX)

Acknowledgments

We thank the colleagues from Interpolis who provided their expertise, support and assistance during the research. We thank Purple, for the development of the IVE and help with the technical details during the experiment. Also we would like to thank the two anonymous reviewers whose comments and questions helped to improve and clarify this manuscript.

Data Availability

All data files are available from the Open Science Framework database. Jansen, P.C.P., Snijders, C.C.P. & Willemsen, M.C. (2020, February 14). Dataset: effect of IVE on prevention behavior. Retrieved from osf.io/kwq45 DOI 10.17605/OSF.IO/KWQ45.

Funding Statement

The first author, Patty Jansen, works three days per week as a marketing researcher at an insurance company (Achmea: https://www.achmea.nl/) and two days per week at Eindhoven University of Technology on her Ph.D. research. Achmea supports her Ph.D. project financially and funded this research. The specific roles of these authors are articulated in the ‘author contributions’ section. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.U.S. Fire Administration. Residential Building Fire Trends (2005–2014) [Internet]. Emmitsburg, Maryland; 2016. Available from: https://www.usfa.fema.gov/downloads/pdf/statistics/res_bldg_fire_estimates.pdf
  • 2.Brandweeracademie. Jaaroverzicht fatale woningbranden 2014 [Internet]. 2015. Available from: http://www.ifv.nl/advieseninnovatie/Documents/20151030-BA-Jaaroverzicht-fatale-woningbranden-2014.pdf
  • 3.Tyler TR, Cook FL. The mass media and judgments of risk: Distinguishing impact on personal and societal level judgments. J Pers Soc Psychol. 1984;47(4):693–708. [Google Scholar]
  • 4.Siegel JM, Shoaf KI, Afifi A a., Bourque LB. Surviving Two Disasters: Does Reaction to the First Predict Response to the Second? Environ Behav. 2003;35(5):637–54. [Google Scholar]
  • 5.Siegrist M, Gutscher H. Natural hazards and motivation for mitigation behavior: People cannot predict the affect evoked by a severe flood. Risk Anal. 2008;28(3):771–8. 10.1111/j.1539-6924.2008.01049.x [DOI] [PubMed] [Google Scholar]
  • 6.Zaalberg R, Midden C, Meijnders A, McCalley T. Prevention, adaptation, and threat denial: Flooding experiences in the Netherlands. Risk Anal. 2009;29(12):1759–78. 10.1111/j.1539-6924.2009.01316.x [DOI] [PubMed] [Google Scholar]
  • 7.Weinstein ND. Effects of personal experience on self-protective behavior. Psychol Bull. 1989;105(1):31–50. 10.1037/0033-2909.105.1.31 [DOI] [PubMed] [Google Scholar]
  • 8.Fogg BJ. Persuasive technology: Using computers to change what we think and do. San Fransisco: Morgan Kauffman; 2003. [Google Scholar]
  • 9.Persky S, Blascovich J. Immersive Virtual Video Game Play and Presence: Influences on Aggressive Feelings and Behavior. Presence Teleoperators Virtual Environ. 2008;17(1):57–72. [Google Scholar]
  • 10.Witmer BG, Singer MJ. Measuring Presence in Virtual Environments: A Presence Questionnaire. Presence Teleoperators Virtual Environ [Internet]. 1998;7(3):225–40. Available from: http://www.mitpressjournals.org/doi/10.1162/105474698565686 [Google Scholar]
  • 11.Peperkorn HM, Diemer J, Mühlberger A. Temporal dynamics in the relation between presence and fear in virtual reality. Comput Human Behav [Internet]. 2015;48:542–7. Available from: 10.1016/j.chb.2015.02.028 [DOI] [Google Scholar]
  • 12.Andrée K, Nilsson D, Eriksson J. Evacuation experiments in a virtual reality high-rise building: Exit choice and waiting time for evacuation elevators. Fire Mater. 2016;40(4):554–67. [Google Scholar]
  • 13.Zaalberg R, Midden CJH. Living Behind Dikes: Mimicking Flooding Experiences. Risk Anal. 2013;33(5):866–76. 10.1111/j.1539-6924.2012.01868.x [DOI] [PubMed] [Google Scholar]
  • 14.Chittaro L, Zangrando N. The Persuasive Power of Virtual Reality: Effects of Simulated Human Distress on Attitudes towards Fire Safety In: Ploug T., Hasle P., Oinas-Kukkonen H, editor. PERSUASIVE 2010. Denmark: Springer; 2010. p. 58–69. [Google Scholar]
  • 15.Chittaro L, Sioni R. Serious games for emergency preparedness: Evaluation of an interactive vs. a non-interactive simulation of a terror attack. Comput Human Behav. 2015;50(September):508–19. [Google Scholar]
  • 16.Chittaro L. Passengers’ Safety in Aircraft Evacuations: Employing Serious Games to Educate and Persuade In: Bang M., Ragnemalm E, editor. PERSUASIVE 2012. Sweden: Springer; 2012. p. 215–26. [Google Scholar]
  • 17.Chittaro L. Changing User’s Safety Locus of Control through Persuasive Play: An Application to Aviation Safety In: A. Spagnolli et al. , editor. PERSUASIVE 2014. Switzerland: Springer; 2014. p. 31–42. [Google Scholar]
  • 18.Chittaro L. Designing serious games for safety education: “Learn to brace” versus traditional pictorials for aircraft passengers. IEEE Trans Vis Comput Graph. 2016;22(5):1527–39. 10.1109/TVCG.2015.2443787 [DOI] [PubMed] [Google Scholar]
  • 19.Chittaro L, Buttussi F. Assessing Knowledge Retention of an Immersive Serious Game vs. a Traditional Education Method in Aviation Safety. IEEE Trans Vis Comput Graph. 2015;21(4):529–38. 10.1109/TVCG.2015.2391853 [DOI] [PubMed] [Google Scholar]
  • 20.Rosenstock IM. Why People Use Health Services. Milbank Mem Fund Q [Internet]. 1966;44(3):94–127. Available from: 10.1111/j.1468-0009.2005.00425.x [DOI] [PubMed] [Google Scholar]
  • 21.Burns AC. The expanded health belief model as a basis for enlightened preventive health care practice and research. J Health Care Mark. 1992;12(3):32–45. [PubMed] [Google Scholar]
  • 22.Janz, Nancy. K., Becker MH. The health belief model: A decade later. Health Educ Q. 1984;11:1–47. [DOI] [PubMed] [Google Scholar]
  • 23.Rosenstock IM, Strecher VJ, Becker MH. Social learning theory and the health belief model. Health Educ Q. 1988;15(2):175–83. 10.1177/109019818801500203 [DOI] [PubMed] [Google Scholar]
  • 24.Rogers RW. A protection motivation theory of fear appeals and attitude change. J Psychol Interdiscip Appl. 1975;91(1):93–114. [DOI] [PubMed] [Google Scholar]
  • 25.Maddux JE, Rogers RW. Protection motivation and self-efficacy: A revised theory of fear appeals and attitude change. J Exp Soc Psychol. 1983;19(5):469–79. [Google Scholar]
  • 26.Ahn SJ, Bailenson JN, Park D. Short- and long-term effects of embodied experiences in immersive virtual environments on environmental locus of control and behavior. Comput Human Behav [Internet]. 2014;39:235–45. Available from: 10.1016/j.chb.2014.07.025 [DOI] [Google Scholar]
  • 27.Proulx G. Occupant behaviour and evacuation. Proc 9th Int Fire Prot Symp. 2001;219–32. [Google Scholar]
  • 28.Hoyt MF. Internal-External Control and Beliefs about Automobile Travel. J Res Pers. 1974;7:288–93. [Google Scholar]
  • 29.Jones JW, Wuebker LJ. Safety Locus Of Control and Employees’ Accidents. J Bus Psychol. 1993;7(4):449–57. [Google Scholar]
  • 30.Wuebker LJ. Safety locus of conrol as a predictor of industrial accidents and injuries. J Bus Psychol. 1986;1(1):19–30. [Google Scholar]
  • 31.Montag I, Comrey AL. Internality and externality as correlates of involvement in fatal driving accidents. J Appl Psychol [Internet]. 1987. August;72(3):339–43. Available from: http://www.ncbi.nlm.nih.gov/pubmed/3624153 [PubMed] [Google Scholar]
  • 32.Chittaro L, Corbett CL, McLean GA, Zangrando N. Safety knowledge transfer through mobile virtual reality: A study of aviation life preserver donning. Saf Sci [Internet]. 2018;102(December 2016):159–68. Available from: 10.1016/j.ssci.2017.10.012 [DOI] [Google Scholar]
  • 33.Manzoni GM, Pagnini F, Gorini A, Preziosa A, Castelnuovo G, Molinari E, et al. Can Relaxation Training Reduce Emotional Eating in Women with Obesity? An Exploratory Study with 3 Months of Follow-Up. J Am Diet Assoc [Internet]. 2009;109(8):1427–32. Available from: 10.1016/j.jada.2009.05.004 [DOI] [PubMed] [Google Scholar]
  • 34.Tarnanas IA, Manos G. A Clinical Protocol for the Development of a Virtual Reality Behavioral Training in Disaster Exposure and Relief In: Wiederhold BK, Riva G, editors. Annual Review of Cybertherapy and Telemedicine 2010 Introduction. San Diego: Interactive Media Institute,; 2004. p. 71–84. [Google Scholar]
  • 35.Armitage CJ. The relationship between multidimensional health locus of control and perceived behavioural control: How are distal perceptions of control related to proximal perceptions of control? Psychol Health. 2003;18(6):723–38. [Google Scholar]
  • 36.Huang JL, Ford JK. Driving locus of control and driving behaviors: Inducing change through driver training. Transp Res Part F Traffic Psychol Behav [Internet]. 2012. May [cited 2013 Oct 11];15(3):358–68. Available from: http://linkinghub.elsevier.com/retrieve/pii/S1369847811000817 [Google Scholar]
  • 37.Schwarzer R, Luszczynska A, Ziegelmann JP, Scholz U, Lippke S. Social-cognitive predictors of physical exercise adherence: three longitudinal studies in rehabilitation. Heal Psychol. 2008;27(1):S54–63. [DOI] [PubMed] [Google Scholar]
  • 38.Martin IM, Bender H, Raish C. What motivates individuals to protect themselves from risks: The case of wildland fires. Risk Anal. 2007;27(4):887–900. 10.1111/j.1539-6924.2007.00930.x [DOI] [PubMed] [Google Scholar]
  • 39.Bradley MM, Greenwald MK, Petry MC, Lang PJ. Remembering Pictures: Pleasure and Arousal in Memory. 1992;18(2):379–90. 10.1037//0278-7393.18.2.379 [DOI] [PubMed] [Google Scholar]
  • 40.Cahill L., McGaugh JL. A novel demonstration of enhanced memory associated with emotional arousal. Conscious Cogn. 1995;4:410–21. 10.1006/ccog.1995.1048 [DOI] [PubMed] [Google Scholar]
  • 41.Kensinger EA, Brierley B, Medford N, Growdon JH, Corkin S. Effects of normal aging and Alzheimer’s disease on emotional memory. Emotion. 2002;2(2):118–34. 10.1037/1528-3542.2.2.118 [DOI] [PubMed] [Google Scholar]
  • 42.Kensinger EA. Remembering the Details: Effects of Emotion. Emot Rev. 2009;1(2):99–113. 10.1177/1754073908100432 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Eck R Van. Digital Game-Based Learning: It’s Not Just the Digital Natives Who Are Restless. Educ Rev. 2006;41(2):1–16. [Google Scholar]
  • 44.Van Eck R. Review of The Ecology of Games: Connecting Youth, Games, and Learning. Educause Quarterly. 2008. 75–78 p. [Google Scholar]
  • 45.Shaffer DW, Squire KR, Halverson R, Gee JP. Video games and the future of learning. Phi delta kappan [Internet]. 2005;87(2):104–11. Available from: http://www.wcer.wisc.edu/publications/workingpapers/working_paper_no_2005_4.pdf [Google Scholar]
  • 46.Gee JP. Learning and games In: Salen K, editor. The Ecology of Games: Connecting Youth, Games, and Learning. Cambridge, MA: The MIT Press; 2008. p. 21–40. [Google Scholar]
  • 47.Kato PM, Cole SW, Bradlyn AS, Pollock BH. A Video Game Improves Behavioral Outcomes in Adolescents and Young Adults With Cancer: A Randomized Trial. Pediatrics [Internet]. 2008;122(2):305–17. Available from: http://pediatrics.aappublications.org/cgi/doi/10.1542/peds.2007-3134 [DOI] [PubMed] [Google Scholar]
  • 48.Wong WL, Shen C, Nocera L, Carriazo E, Tang F, Bugga S, et al. Serious Video Game Effectiveness. Proc Int Conf Adv Comput Entertain Technol [Internet]. 2007;49–55. Available from: http://portal.acm.org/citation.cfm?doid=1255047.1255057 [Google Scholar]
  • 49.Rotter JB. Social learning and clinical psychology Englewood Cliffs, N.J.: Prentice-Hall; 1954. [Google Scholar]
  • 50.Suárez-Álvarez J, Pedrosa I, García-Cueto E, Muñiz J. Locus of Control revisited: Development of a new bi-dimensional measure. An Psicol [Internet]. 2016;32(2):578 Available from: http://revistas.um.es/analesps/article/view/analesps.32.2.200781 [Google Scholar]
  • 51.Hunter DR. Development of an Aviation Safety Locus of Control Scale. Aviat Space Environ Med. 2002;00(0):1–5. [PubMed] [Google Scholar]
  • 52.Parkes KR. Dimesionality of Rotter’s locus of control scale: an application of the “very simple structure” technique. Pers Individ Dif. 1985;6(1):115–9. [Google Scholar]
  • 53.Rotter JB. Generalized expectancies for internal versus external control of reinforcement. Psychol Monogr Gen Appl. 1966;80(1):1–28. [PubMed] [Google Scholar]
  • 54.Murray CD, Fox J, Pettifer S. Absorption, dissociation, locus of control and presence in virtual reality. Comput Human Behav. 2007;23(3):1347–54. [Google Scholar]
  • 55.Schwebel DC, McClure LA, Porter BE. Experiential exposure to texting and walking in virtual reality: A randomized trial to reduce distracted pedestrian behavior. Accid Anal Prev [Internet]. 2017;102:116–22. Available from: 10.1016/j.aap.2017.02.026 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 56.Luszczynska A, Schwarzer R. Multidimensional Health Locus of Control: Comments on the construct and its measurement. J Health Psychol. 2005;10(5):633–42. 10.1177/1359105305055307 [DOI] [PubMed] [Google Scholar]
  • 57.Bandura A. Self-Efficacy. Encycl Hum Behav. 1994;4(1994):71–81. [Google Scholar]
  • 58.Lieberman DA. Interactive video games for health promotion: Effects on knowledge, self-efficacy, social support and health In: Street R.L., Gold W.R., & Manning T, editor. Health promotion and interactive technology: theoretical applications and future directions. Mahwah, NJ: Lawrence Erlbaum Associates; 1997. [Google Scholar]
  • 59.Lieberman DA. Health Education Video Games for Children and Adolescents: Theory, Design, and Research Findings. 1998; Available from: http://www.eric.ed.gov/ERICWebPortal/detail?accno=ED422312 [Google Scholar]
  • 60.Svenson O. Are we all less risky and more skillful than our fellow drivers? Acta Psychol (Amst). 1981;47(2):143–8. [Google Scholar]
  • 61.Sümer N, Özkan T, Lajunen T. Asymmetric relationship between driving and safety skills. Accid Anal Prev. 2006;38(4):703–11. 10.1016/j.aap.2005.12.016 [DOI] [PubMed] [Google Scholar]
  • 62.Taubman-Ben-Ari O, Mikulincer M, Iram A. A multi-factorial framework for understanding reckless driving-appraisal indicators and perceived environmental determinants. Transp Res Part F Traffic Psychol Behav. 2004;7(6):333–49. [Google Scholar]
  • 63.Morisset N, Terrade F, Somat A. Perceived self-efficacy and risky driving behaviors: The mediating role of subjective risk judgment. Swiss J Psychol. 2010;69(4):233–8. [Google Scholar]
  • 64.Proulx G. Playing with fire: why people linger in burning buildings. ASHRAE J. 2003;45(7):33–5. [Google Scholar]
  • 65.Amir M, Roziner I, Knoll a, Neufeld MY. Self-efficacy and social support as mediators in the relation between disease severity and quality of life in patients with epilepsy. Epilepsia. 1999;40(2):216–24. 10.1111/j.1528-1157.1999.tb02078.x [DOI] [PubMed] [Google Scholar]
  • 66.Somers TJ, Shelby RA, Keefe FJ, Godiwala N, Lumley MA, Mosley-Williams A, et al. Disease severity and domain-specific arthritis self-efficacy: Relationships to pain and functioning in patients with rheumatoid arthritis. Arthritis Care Res. 2010;62(6):848–56. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 67.Wallston BS, Wallston KA, Kaplan GD, Maides SA. Development and validation of the health locus of control (HLC) scale. J Consult Clin Psychol. 1976;44(4):580–5. 10.1037//0022-006x.44.4.580 [DOI] [PubMed] [Google Scholar]
  • 68.de Hoog N, Stroebe W, de Wit JBF. The processing of fear-arousing communications: How biased processing leads to persuasion. Soc Influ [Internet]. 2008. June [cited 2015 Aug 10];3(2):84–113. Available from: http://www.tandfonline.com/doi/abs/10.1080/15534510802185836 [Google Scholar]
  • 69.Schwarzer R, Jerusalem M. Generalized Self-Efficacy scale In: Weinman J., Wright S., Johnston M, editor. Measures in health psychology: A user’s portfolio Causal and control beliefs. Windsor, UK: Nfer-Nelson; 1995. p. 35–7. [Google Scholar]
  • 70.Bandura A. Guide for constructing self-efficacy scales. Self-efficacy beliefs Adolesc. 2006;307–37. [Google Scholar]
  • 71.Weber S. Bacon: An effective way to detect outliers in multivariate data using Stata (and Mata). Stata J. 2010;10(3):331–8. [Google Scholar]
  • 72.D’agostino RB, Belanger A, D’Agostino RB Jr. A Suggestion for Using Powerful and Informative Tests of Normality. Am Stat. 1990;44(4):316–21. [Google Scholar]
  • 73.Shapiro SS, Martin BW. An Analysis of Variance Test for Normality. Biometrika. 1965;52(3/4):591–611. [Google Scholar]
  • 74.Kline RB. Principles and practice of structural equation modeling 4th ed. New York: The Guilford Press; 2016. [Google Scholar]
  • 75.Muthén L., Muthén B. Mplus Version 8. Los Angeles: Muthén, & Muthén; [Google Scholar]
  • 76.Bandalos DL. Relative Performance of Categorical Diagonally Weighted Least Squares and Robust Maximum Likelihood Estimation. Struct Equ Model. 2014;21(1):102–16. [Google Scholar]
  • 77.Hair JF, Black WC, Babin BJ, Anderson RE. Multivariate Data Analysis. 7th ed. Upper Saddle River: NJ: Pearson Prentice Hall; 2010. [Google Scholar]
  • 78.Browne MW, Cudeck R. Alternative ways of assessing model fit In: Bollen K.A, Long JS, editor. Testing structural equation models. Thousand Oaks, CA: Sage; 1993. p. 136–62. [Google Scholar]
  • 79.Hu L, Bentler PM. Cutoff criteria for fit indexes in covariance structure analysis: Conventional criteria versus new alternatives. Struct Equ Model A Multidiscip J. 1999;6(1):1–55. [Google Scholar]
  • 80.Tucker LR, Lewis C. A reliability coefficient for maximum likelihood factor analysis. Psychometrika. 1973;38(1):1–10. [Google Scholar]
  • 81.Knijnenburg BP, Willemsen MC. Evaluating Recommender Systems with User Experiments In: Ricci F, Rokach L., Shapira B, editor. Recommender Systems Handbook [Internet]. New York: Springer Science+Business Media; 2015. p. 309–52. Available from: http://www.springerlink.com/index/10.1007/978-0-387-85820-3 [Google Scholar]
  • 82.Brown TA. Methodology in the social sciences Confirmatory factor analysis for applied research. Kenny DA, editor. New York: The Guilford Press; 2006. 475 p. [Google Scholar]
  • 83.Ryon HS, Gleason MEJ. The Role of Locus of Control in Daily Life. Personal Soc Psychol Bull. 2014;40(1):121–31. [DOI] [PubMed] [Google Scholar]
  • 84.Bubeck P, Botzen WJW, Aerts JCJH. A Review of Risk Perceptions and Other Factors that Influence Flood Mitigation Behavior. Risk Anal [Internet]. 2012;32(9):1481–95. Available from: 10.1111/j.1539-6924.2011.01783.x [DOI] [PubMed] [Google Scholar]
  • 85.Shrout PE, Bolger N. Mediation in experimental and nonexperimental studies: New procedures and recommendations. Psychol Methods. 2002;7(4):422–45. [PubMed] [Google Scholar]
  • 86.Zaalberg R, Midden C. Enhancing Human Responses to Climate Change Risks In: Ploug T., Hasle P., Oinas-Kukkonen H, editor. PERSUASIVE 2010. Denmark: Springer; 2010. p. 205–10. [Google Scholar]

Decision Letter 0

Geilson Lima Santana

19 Aug 2019

PONE-D-19-17627

Playing with fire. Understanding how experiencing a fire in an immersive virtual environment affects prevention behavior.

PLOS ONE

Dear Ms. Jansen,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

We would appreciate receiving your revised manuscript by Oct 03 2019 11:59PM. When you are ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter.

To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). This letter should be uploaded as separate file and labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. This file should be uploaded as separate file and labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. This file should be uploaded as separate file and labeled 'Manuscript'.

Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out.

We look forward to receiving your revised manuscript.

Kind regards,

Geilson Lima Santana, M.D., Ph.D.

Academic Editor

PLOS ONE

Journal Requirements:

1. When submitting your revision, we need you to address these additional requirements.

Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

http://www.journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and http://www.journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. We note that you have stated that you will provide repository information for your data at acceptance. Should your manuscript be accepted for publication, we will hold it until you provide the relevant accession numbers or DOIs necessary to access your data. If you wish to make changes to your Data Availability statement, please describe these changes in your cover letter and we will update your Data Availability statement to reflect the information you provide.

3. Thank you for stating the following in the Financial Disclosure section:

In the interest of full disclosure, we wish to draw your attention to the following: the first author, Patty Jansen, works three days per week as a marketing researcher at an insurance company (Achmea) and two days per week at Eindhoven University of Technology on her Ph.D. research. Her employer contributes partly to her Ph.D. research by financially supporting her for one day per week.  The employer also financially supported this research. We declare that the research that we present was in no way influenced by the insurance company. The funder had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

https://www.achmea.nl/

We note that you received funding from a commercial source: Achmea

Please provide an amended Competing Interests Statement that explicitly states this commercial funder, along with any other relevant declarations relating to employment, consultancy, patents, products in development, marketed products, etc.

Within this Competing Interests Statement, please confirm that this does not alter your adherence to all PLOS ONE policies on sharing data and materials by including the following statement: "This does not alter our adherence to PLOS ONE policies on sharing data and materials.” (as detailed online in our guide for authors http://journals.plos.org/plosone/s/competing-interests).  If there are restrictions on sharing of data and/or materials, please state these. Please note that we cannot proceed with consideration of your article until this information has been declared.

Please include your amended Competing Interests Statement within your cover letter. We will change the online submission form on your behalf.

Please know it is PLOS ONE policy for corresponding authors to declare, on behalf of all authors, all potential competing interests for the purposes of transparency. PLOS defines a competing interest as anything that interferes with, or could reasonably be perceived as interfering with, the full and objective presentation, peer review, editorial decision-making, or publication of research or non-research articles submitted to one of the journals. Competing interests can be financial or non-financial, professional, or personal. Competing interests can arise in relationship to an organization or another person. Please follow this link to our website for more details on competing interests: http://journals.plos.org/plosone/s/competing-interests

4. "We note from your Cover Letter and Financial Disclosure that the first author of the manuscript (FCPJ) is affiliated to an insurance company. Please update the affiliations within your Cover Page and within Editorial Manager to reflect this affiliation.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Partly

Reviewer #2: Partly

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: No

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: No

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: I’ve read the paper with interest. It’s about a timely topic (comparing VR and non-VR approaches for teaching safety behaviors) and aims at exploring an additional variable with respect to other studies.

The way the paper is written does not yet meet the requirements of an high-quality presentation, so I provide a number of comments in the following about necessary improvements, that I hope we will be helpful in preparing a thoroughly revised version.

First, the paper needs to tone down its initial claims of extreme originality (three of the four claims are exaggerated).

“Our study has four characteristics that, taken together, make it stand out from most previous IVE studies.” COMMENT: Actually this claim is not accurate and must be toned down for three of the four characteristics.

“First, our study focuses on the fire domain, while most risk related IVE-effect studies

involve other risk domains (e.g. health, traffic safety, environmental risk, aircraft evacuation).” COMMENT: Several IVEs concerning fires have actually been built, see this scholar search as a starting point:

https://scholar.google.com/scholar?hl=en&as_sdt=0%2C5&q=%22virtual+reality%22+Fire&btnG=

“Second, we consider the effects of IVE on a full set of “psychological determinants

(knowledge, vulnerability, severity, self-efficacy, and locus of control). “ COMMENT: Also effects of virtual risk experiences on such variables have been previously studied (and the paper should contrast its research with the other findings on these psychological variables), for example:

Chittaro L., Designing Serious Games for Safety Education: "Learn to Brace" vs. Traditional Pictorials for Aircraft Passengers, IEEE Transactions on Visualization and Computer Graphics, 22, 2016, 1527-1539.

Chittaro L., Sioni R., Serious Games for Emergency Preparedness: Evaluation of an Interactive vs. a Non-Interactive Simulation of a Terror Attack, Computers in Human Behavior, 50, 2015, 508–519.

Both these papers also refer to the same theoretical basis of the submitted manuscript (Protection Motivation Theory, PMT)

“Third, we compare the effects of IVE with a control condition which allows establishing the effect of IVE over and above a more standard way of getting the same information across.” COMMENT: See above

“Fourth, and most importantly, we measure actual prevention behavior, and test to what extent psychological determinants have an effect on actual behavior.” COMMENT: This is the actual key point to stress. It’s the characteristic on which this research stands out.

However, the authors have to be careful about terminology concerning behavior. The two behaviors measured in the study were 1) if a participant would invest part of his or her show-up fee in a fire blanket and 2) whether a participant would take home flyers related to fire safety. There’s a terminological issue all over the paper (that can however be easily fixed): the paper uses the term “prevention behaviors” to refer to the two behaviors, but what they actually show is a possible interest in prevention (for example, taking home a flyer is no guarantee that the participant will keep and read the flyer and take the preventive actions described in the flyer).

The paper (correctly) stresses the importance of VR presence creating the expectation that the construct will be measured in the study, but then it is not. This is a bit confusing.

According to PMT, a simple hypothesis that Perceived Vulnerability and Perceived Severity positively influence behavior, as the paper does (H3b and H4b) is not straightforwardly supported. The theory indeed clearly states that if vulnerability and severity are perceived, but the recommended behavior is not presented in a way that convinces the message recipient about efficacy (both recommendation efficacy and self-efficacy), then the influence will be negative instead of positive: the recipient will engage in emotion control behaviors instead of the recommended risk control (prevention) behavior.

Hypothesis “H5c: Perceived self-efficacy negatively influences prevention behavior” needs more theoretical motivation. As it is, it apparently contradicts the theory (PMT) on which the paper aims at being substantially grounded.

The description of the IVE condition is lacking of enough details. For example, the paper says “The participant can then choose between different actions: 1.) go towards the source of the fire (the kitchen) and extinguish the fire with the fire blanket. The fire will be extinguished and the game ends. Or 2.) go towards…”

A first question is how such actions are selected? Second, the granularity of the actions is not clear. For example, what happens if one selects action 1? All those listed things happen as the user passively watches? Or does the user have to perform them? In this latter case, there are multiple actions. And what does the user do in the physical world to perform the multiple actions?

More generally, since one of the goals of the IVE in the paper is to increase users’ knowledge, it is important to fully describe everything users could do and what was the feedback from the environment, also highlighting the pedagogical aspects of the IVE design.

Attaching videos to the submission could provide some additional help.

In the hardware section, the paper says the software was implemented on one

desktop machine and one laptop. But what happened during the study? Some people used the laptop and some other the desktop?

The measurement section is not always clear too. The first measurement is said to “include current prevention behaviors that are beyond the score (sic) of this paper.” What does this mean? Were people asked about their fire prevention behaviors?

Then the paper says “The prevention behaviors that were measured in the first and third measurement are not analyzed in this paper”. One can try to guess that this was done to assess if there was a behavior change (at least a self-reported ones). If this is the case, it’s strange that the paper omits to analyze it, since the main difference from other research in the literature concerns specifically participant’s behavior.

To measure the SLOC construct, the paper creates a questionnaire in which some items are supposed to measure internal SLOC while the others external SLOC. However, the section does not report a factor analysis to confirm that the set of created items can actually be divided in the two subscales.

One of the items (“How much panic do you think there will be in case of a grease fire?”) that were used to measure severity is different from the usual items that measure that construct, and the choice should be explained.

One of the items (“People should be rewarded by their insurance company if they take

preventive measures to prevent or control fire”) that were used to measure SLOC does not actually concern locus of control orientation.

Unlike the first two questionnaires, the third questionnaire was administered on-line. The paper does not provide details about actions taken to prevent issues of on-line surveys, for example: were the answering times measured and taken care of ?

The “descriptive statistics” says that the effect of the IVE is smaller than expected, but provides no effect sizes in the statistics.

In Table 1, the translations from Dutch to English is not always good, in one case “Fire in a home usually exists by the people themselves” the reader cannot be sure of what that is supposed to mean.

The percentage of participants who felt nauseous during the VR experience is surprisingly high and needs discussion. It’s an indication that some VR design guidelines were likely violated by the implemented IVE. This can limit significantly the effectiveness of the IVE for the goal intended in the paper.

The result that participants using VR learned less than those using printed information needs much more discussion. From the details and screenshots provided, a simple and likely explanation can be traced back to the design and implementation of the IVE, which seems too primitive compared with the state of the art in the VR literature. The paper should discuss the differences in interactivity, realism, and pedagogical methods between the IVE used and the more sophisticated IVEs available in the literature.

Moreover, since the IVE can influence psychological determinants both indirectly (through level of knowledge) and directly, the limitations in the IVE can have had consequences that reach furthest than knowledge only, and weaken the conclusions of the paper. This should be discussed in a critical way.

Care must also be taken in the conclusions section which currently seems to assume that the results hold for IVEs in general, while the limitations in the studied IVE make it difficult to support generalization to more complex IVEs used in learning and training.

Reviewer #2: In the present manuscript, the authors provide an interesting approach towards fire prevention by using an immersive virtual environment procedure to teach people how to react in the case of a grease fire. The inclusion of actual prevention behavior, together with self-report measures and the use of an experimental setting with randomization and control condition gives this article a promising potential. However, there are some writing and data-analytic issues that, in my opinion, need to be addressed before a recommendation for publication. In my review, I will focus mostly in consistency, writing form and methods which is my area of expertise.

The statistical analyses (SEM) done in this article do not take full advantage of the experimental setting created by the authors, given that the results are not controlled by baseline levels (as in a group x time interaction model), which is a problem the authors should address in detail. It is a common practice to measure the constructs in a pre-post fashion, however the authors did not do it to avoid priming. However, the trade-off is that there is no baseline control anymore, turning the analyses correlational, which should be addressed by the authors in more depth.

On the other hand, I find the data analysis and results section rather confusing. This is, in my opinion, mostly because it is very data dense and contains a lot of modelling, which always creates a challenge in reporting. One consequence of this, is the omission of important information to evaluate the quality of the data analyses. This is especially relevant for the CFA and SEM parts, where they show some inconsistencies that I’m listing below.

pp 6 (and others). When talking about psychological determinants, please be more specific. Psychological determinants of prevention behavior?

pp 7: "The underlying arguments are equally appropriate for the area of fire prevention behavior" It is important to give arguments when presenting such a statement, why?

Pp 7. "Studies usually use" is not the best argument for selection of variables: I strongly suggest the authors to describe a sound logical and empirical basis for their proposal. This is done later; however, it can be briefly explicated here.

Pp.7 "We feel it is appropriate or at least worthwhile to study psychological determinants and the target behaviors simultaneously" Why? It is, indeed, important, but this should be elaborated.

pp 8. "Can in fact be studied empirically relatively easily" How?

pp11. Please check repeated citation (29,29).

pp11. Please provide references for the SLOC bi-dimentionality, if they were provided before, I suggest moving the definition of the construct together with the citations above.

pp 22. Why are only knowledge, vulnerability and severity included in the third measurement? This exclusion should be supported.

Measures: Please report basic psychometric properties of the instruments here (cronbach's alpha for internal consistency, for example, and factor structure of original instruments).

pp. 27. Please specify the software version used for SEM analyses.

Data analyses and report:

Please describe which sampling wave was used for the analyses (I guess it’s sample 2, however I cannot find it in the manuscript).

Measurement models: I find it problematic to test all measurement models together, given that this would further obscure local sources of bad fit. Is it just one instrument the one creating the bad fit? Are all measurement models wrong? I would suggest test this separately. On the other hand, the procedure of co-variation between items within an instrument should be done in this step.

Structural Equation Models: Modifications were made to the models, which is a step towards a good fit. It is important to see if the modifications suggested by the modification indexes are reasonable under theoretical grounds. This is done for the co-variation between variables; however, it should also be made for the covariation between items. It is usually accepted to co-vary items within a scale if they share method or another known source of shared variance that should not be captured by the latent factor (please see previous comment, however, regarding the location of this procedure).

It is not clear why do the authors base their report on model 4, if the model with a good fit is model 6. It is important to report the good fitting model. Even though models showed no substantial changes in their point estimates, the non-significant results are part of a bad-fitting model and thus are not directly interpretable (i.e. SLOC). This should be at least stated as a limitation or tested with a different regression technique.

I find it important to explicitly say, on the SEM models, how were the dichotomous outcomes treated, and interpret them accordingly. Where they estimated using a probit or logistic regression? (In Mplus this is different for different estimation methods)? Please clarify this on the manuscript.

Please also describe which method did the authors use to compute the mediation analyses (multiplicative?) and provide the estimates in the tables.

Pp 36. Please specify which bootstrap method was used. Bias corrected, percentile? They have shown different trade-offs regarding type 1 errors.

Pp 36. A direct effect of IVE on fire blanket is reported, however later it is exposed that no significant effect of IVE on behavior was found. Please clarify this.

I think here the problem may arise because the total effect is non-significant while the direct effect is significant when the mediator is partialized. By checking the estimates, it is possible to see, as the authors indicated, that indirect and direct effects have opposite signs. This, together with the fact that the direct effect is bigger than the total effect is an indication of an inconsistent mediation, where the mediator acts as a suppressor variable in the relationship between IVE and blanket. Given that the mediator is part of the model, this may be interpreted as a significant direct effect when the mediator is taken into account, and should be discussed accordingy.

Pp 36. A nonsignificant effect is presented as "marginally significant". I suggest avoiding using this language on the manuscript because they can be misleading.

Pp 37. Please provide assumption checks for the repeated measures anova, and report results in a corresponding table.

Discussion:

The authors propose here that the relationship between IVE and self-efficacy is partly mediated by vulnerability, however in their model (figure 3) they estimate a co-variation and not an indirect effect properly. Please clarify this point, and in the case there is an indirect effect, please report it accordingly. If no mediation effect was computed, this should be not discussed as such.

Please provide a substantial interpretation of the significant mediation effect of IVE to blanket towards knowledge, so the reader can get an idea of the mechanism involved.

The authors discuss the nonsignificant interaction effect. In this case, it is advisable to avoid discussing nonsignificant results as “marginally significant”. The authors describe this as a “fully” mediated effect, which may not be the case given what I exposed before.

The authors discuss the importance of causal claims; however, this is beyond the scope of the present article (I describe this at the beginning). Same for the influence of IVE on the psychological domains, it is important to control for previous values.

General comments:

The authors usually justify their decisions with statements such as "We feel it is appropriate or at least worthwhile" (pp 7), "we felt it was relevant to compare the IVE experience with an information sheet". Although their decisions seem reasonable from a methodological point of view, they need to be justified based on methodological or theoretical grounds (i.e. benefit of simultaneous measurement of self-report and behavior, benefit of a control condition versus no control condition).

Please state the difference between serious and commercial games.

I hope the authors find this review useful for their research!

Best regards,

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: Yes: Cristóbal Hernández C.

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files to be viewed.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2020 Mar 6;15(3):e0229197. doi: 10.1371/journal.pone.0229197.r002

Author response to Decision Letter 0


30 Oct 2019

PONE-D-19-17627

Playing with fire. Understanding how experiencing a fire in an immersive virtual environment affects prevention behavior.

PLOS ONE

Dear Ms. Jansen,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

We would appreciate receiving your revised manuscript by Oct 03 2019 11:59PM. When you are ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter.

To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

Please include the following items when submitting your revised manuscript:

• A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). This letter should be uploaded as separate file and labeled 'Response to Reviewers'.

• A marked-up copy of your manuscript that highlights changes made to the original version. This file should be uploaded as separate file and labeled 'Revised Manuscript with Track Changes'.

• An unmarked version of your revised paper without tracked changes. This file should be uploaded as separate file and labeled 'Manuscript'.

Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out.

We look forward to receiving your revised manuscript.

Kind regards,

Geilson Lima Santana, M.D., Ph.D.

Academic Editor

PLOS ONE

Journal Requirements:

1. When submitting your revision, we need you to address these additional requirements.

Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

http://www.journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and http://www.journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

For this resubmission we converted the figure files via PACE.

For the rest, we thought that we adhered to all PLOS ONE’s style requirements. If we missed something, can you please point us to the specific style criteria that we have to adjust?

2. We note that you have stated that you will provide repository information for your data at acceptance. Should your manuscript be accepted for publication, we will hold it until you provide the relevant accession numbers or DOIs necessary to access your data. If you wish to make changes to your Data Availability statement, please describe these changes in your cover letter and we will update your Data Availability statement to reflect the information you provide.

OK

3. Thank you for stating the following in the Financial Disclosure section:

In the interest of full disclosure, we wish to draw your attention to the following: the first author, Patty Jansen, works three days per week as a marketing researcher at an insurance company (Achmea) and two days per week at Eindhoven University of Technology on her Ph.D. research. Her employer contributes partly to her Ph.D. research by financially supporting her for one day per week. The employer also financially supported this research. We declare that the research that we present was in no way influenced by the insurance company. The funder had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

https://www.achmea.nl/

We note that you received funding from a commercial source: Achmea

Please provide an amended Competing Interests Statement that explicitly states this commercial funder, along with any other relevant declarations relating to employment, consultancy, patents, products in development, marketed products, etc.

Within this Competing Interests Statement, please confirm that this does not alter your adherence to all PLOS ONE policies on sharing data and materials by including the following statement: "This does not alter our adherence to PLOS ONE policies on sharing data and materials.” (as detailed online in our guide for authors http://journals.plos.org/plosone/s/competing-interests). If there are restrictions on sharing of data and/or materials, please state these. Please note that we cannot proceed with consideration of your article until this information has been declared.

Please include your amended Competing Interests Statement within your cover letter. We will change the online submission form on your behalf.

Please know it is PLOS ONE policy for corresponding authors to declare, on behalf of all authors, all potential competing interests for the purposes of transparency. PLOS defines a competing interest as anything that interferes with, or could reasonably be perceived as interfering with, the full and objective presentation, peer review, editorial decision-making, or publication of research or non-research articles submitted to one of the journals. Competing interests can be financial or non-financial, professional, or personal. Competing interests can arise in relationship to an organization or another person. Please follow this link to our website for more details on competing interests: http://journals.plos.org/plosone/s/competing-interests

We included a Competing Interests Statements.

4. "We note from your Cover Letter and Financial Disclosure that the first author of the manuscript (FCPJ) is affiliated to an insurance company. Please update the affiliations within your Cover Page and within Editorial Manager to reflect this affiliation.

We included the affiliation within the Cover Page.

We do however not understand how the information at “current address” differs from the “affiliation” information if postal codes or street addresses are not allowed. Also, we were not totally sure if we correctly used the provided symbols. If we made any mistake here, please let us know so that we can correct it.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Partly

Reviewer #2: Partly

________________________________________

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: No

________________________________________

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: No

Reviewer #2: Yes

________________________________________

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

________________________________________

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

We thank both reviewers for the constructive comments, and have chosen to implement almost all of them. We copied and numbered the reviewers’ questions and wrote our answers directly below the questions of the reviewer. The page numbers correspond with the file ‘Manuscript’.

Reviewer #1: I’ve read the paper with interest. It’s about a timely topic (comparing VR and non-VR approaches for teaching safety behaviors) and aims at exploring an additional variable with respect to other studies.

The way the paper is written does not yet meet the requirements of an high-quality presentation, so I provide a number of comments in the following about necessary improvements, that I hope we will be helpful in preparing a thoroughly revised version.

First, the paper needs to tone down its initial claims of extreme originality (three of the four claims are exaggerated).

“Our study has four characteristics that, taken together, make it stand out from most previous IVE studies.” COMMENT: Actually this claim is not accurate and must be toned down for three of the four characteristics.

1. “First, our study focuses on the fire domain, while most risk related IVE-effect studies

involve other risk domains (e.g. health, traffic safety, environmental risk, aircraft evacuation).” COMMENT: Several IVEs concerning fires have actually been built, see this scholar search as a starting point:

https://scholar.google.com/scholar?hl=en&as_sdt=0%2C5&q=%22virtual+reality%22+Fire&btnG=

“Second, we consider the effects of IVE on a full set of “psychological determinants

(knowledge, vulnerability, severity, self-efficacy, and locus of control). “ COMMENT: Also effects of virtual risk experiences on such variables have been previously studied (and the paper should contrast its research with the other findings on these psychological variables), for example:

Chittaro L., Designing Serious Games for Safety Education: "Learn to Brace" vs. Traditional Pictorials for Aircraft Passengers, IEEE Transactions on Visualization and Computer Graphics, 22, 2016, 1527-1539.

Chittaro L., Sioni R., Serious Games for Emergency Preparedness: Evaluation of an Interactive vs. a Non-Interactive Simulation of a Terror Attack, Computers in Human Behavior, 50, 2015, 508–519.

Both these papers also refer to the same theoretical basis of the submitted manuscript (Protection Motivation Theory, PMT)

“Third, we compare the effects of IVE with a control condition which allows establishing the effect of IVE over and above a more standard way of getting the same information across.” COMMENT: See above

“Fourth, and most importantly, we measure actual prevention behavior, and test to what extent psychological determinants have an effect on actual behavior.” COMMENT: This is the actual key point to stress. It’s the characteristic on which this research stands out.

Response to Reviewer comment No. 1:

Just for argument’s sake, we would like to mention that our statement “Our study has four characteristics that, taken together, make it stand out from most previous IVE studies.”, is actually correct if taken to mean that there aren’t that many studies that take all four characteristics into account. We agree with the reviewer though, in the sense that it can easily be read as meaning to say that each of them separately is new.

We toned down the text and changed our text as follows. First, we have removed the argument of the control group as indeed several more IVE-studies have used a non-interactive control. We think the other two arguments still hold, but need more explanation, which we have now added (see p. 3).

[Application to fire prevention]: Several IVE’s concerning fires have been built, but they typically are used to study human behavior in the case of a fire or they serve as training purposes for fire men or to teach fire drill skills to children. What we meant to say is that this is the first IVE-effect study in the fire domain, in which the IVE serves as a means (through the simulation of fire event) to induce changes in the psychological determinants, and subsequently change behavior. We have now added this argument to the text (p. 3).

[Full set of psychological determinants]: Previous studies have indeed studied these variables before, and we also refer to studies of Chittaro in our paper. We were however not aware of the papers that the reviewer mentioned, and thank the reviewer for the suggestions. We have now integrated the given references in our paper (see ref 15 and 18). What none of these studies do, however, is analyze the effects of the IVE on these variables simultaneously. This is one beneficial feature of our Structural Equation Modelling approach. We now elaborate on this in the text (p. 3).

2. However, the authors have to be careful about terminology concerning behavior. The two behaviors measured in the study were 1) if a participant would invest part of his or her show-up fee in a fire blanket and 2) whether a participant would take home flyers related to fire safety. There’s a terminological issue all over the paper (that can however be easily fixed): the paper uses the term “prevention behaviors” to refer to the two behaviors, but what they actually show is a possible interest in prevention (for example, taking home a flyer is no guarantee that the participant will keep and read the flyer and take the preventive actions described in the flyer).

Response to Reviewer comment No. 2:

We partly agree with the reviewer here. Investment of the show-up fee in the fire blanket is an observed behavior as well as taking home flyers. We refer to it as behavior in contrast to the more standard form of self-reported behavior (e.g. “I have the intention to buy a fire blanket”). We believe the term “possible interest”, as suggested by the reviewer, does not really cover the purchase of a fire blanket, which is a real prevention measure. But we agree that a flyer in itself does not reduce the probability nor severity of a fire. This was actually already noted in our Method section: “The primary measurement for the effect of IVE on prevention behavior will be the participants’ purchase of a fire blanket, and the secondary measurement is the participants’ interest in fire prevention information.” (p. 25). We have added a remark about the taking home of the flyer being a rather ‘soft’ kind of behavior in the text on page 25.

3. The paper (correctly) stresses the importance of VR presence creating the expectation that the construct will be measured in the study, but then it is not. This is a bit confusing.

Response to Reviewer comment No. 3:

We do not think the concept of presence was relevant in this study for comparing the conditions, as we compared an IVE with an INFO condition. The INFO condition was not expected to have any effect on “presence” and questions about the level of presence concerning the information sheet would have been confusing. For evaluation purposes we did however include one item after the IVE about the level of realism of the IVE (see p. 19, p. 27).

4. According to PMT, a simple hypothesis that Perceived Vulnerability and Perceived Severity positively influence behavior, as the paper does (H3b and H4b) is not straightforwardly supported. The theory indeed clearly states that if vulnerability and severity are perceived, but the recommended behavior is not presented in a way that convinces the message recipient about efficacy (both recommendation efficacy and self-efficacy), then the influence will be negative instead of positive: the recipient will engage in emotion control behaviors instead of the recommended risk control (prevention) behavior.

Response to Reviewer comment No. 4:

We agree with the reviewer that the behavior -owning a fire blanket- must be perceived as effective or useful in order to be performed. The IVE stresses the response efficacy of the fire blanket, as the viewer observes the fire being extinguished by the use of the fire blanket, but we did not measure the perceived response efficacy. Although this could have been a useful addition (as mentioned in the limitation section on p. 44), we did not want to prime participants in the direction of a fire blanket by asking specific questions about the fire blanket. We added this explanation on p. 44.

5. Hypothesis “H5c: Perceived self-efficacy negatively influences prevention behavior” needs more theoretical motivation. As it is, it apparently contradicts the theory (PMT) on which the paper aims at being substantially grounded.

Response to Reviewer comment No. 5:

Self-efficacy in our paper refers to the confidence people have in their own capacity to correctly handle the situation of a grease fire without having a fire blanket (only participants without a fire blanket were recruited). People often downplay the severity of a fire, overestimate the time they have to handle the fire, and overestimate their own capabilities (e.g. remain calm and use the lid of a pan). This might result in taking less prevention measures.

This is comparable to the road safety domain, in which a higher level of self-efficacy (over estimation of one’s driving capabilities) is related to more unsafe behaviors and accidents. It is common in this domain to develop interventions to decrease someone’s perceived self-efficacy in order to enhance safe behaviors. We integrated several additional references in the paper that show the relationship between self-efficacy and unsafe driving behaviors (p. 15).

An alternative that we rejected, but perhaps would have been closer to other applications of PMT, would have been to try to increase self-efficacy as defined by someone’s confidence to be able to extinguish the grease fire with a fire blanket. This would have been more appropriate when we would have been aiming at explaining the actual use of a fire blanket, instead of buying it. In addition, in the current IVE the participant just had to focus their eyes a couple of seconds to the fire blanket to “extinguish the fire with the fire blanket”. To really increase self-efficacy related to using the fire blanket in the IVE, special gloves should ideally be used to realistically simulate gestures of picking up the blanket and putting it on the grease fire. To show the participant how it should be done, and increase one’s self-confidence in using the fire blanket successfully.

Because we wanted to focus on the purchase of the fire blanket, we chose to try and decrease the level of self-efficacy (overestimation of correctly handle a grease fire situation), as is common in the road safety domain, in order to stimulate the need for extra prevention measures (as mentioned on p. 15, 16).

6. The description of the IVE condition is lacking of enough details. For example, the paper says “The participant can then choose between different actions: 1.) go towards the source of the fire (the kitchen) and extinguish the fire with the fire blanket. The fire will be extinguished and the game ends. Or 2.) go towards…”

A first question is how such actions are selected? Second, the granularity of the actions is not clear. For example, what happens if one selects action 1? All those listed things happen as the user passively watches? Or does the user have to perform them? In this latter case, there are multiple actions. And what does the user do in the physical world to perform the multiple actions?

More generally, since one of the goals of the IVE in the paper is to increase users’ knowledge, it is important to fully describe everything users could do and what was the feedback from the environment, also highlighting the pedagogical aspects of the IVE design.

Attaching videos to the submission could provide some additional help.

Response to Reviewer comment No. 6:

We agree with the reviewer that we could have been clearer about the IVE and the possible scenarios. We have now included more details in our text and added a detailed description of each scenario in the S2 file (p. 20-22).

The participant can move through the environment (with a joystick) and can perform different actions by focusing on objects with his eyes (e.g. open doors, pick up a toddler, use fire blanket). So the participant does not passively watch, but does everything him / herself, albeit with relative simple controls.

On YouTube one can see a video of a similar IVE, a version made that was made for the Samsung Gear (in which a smartphone has to be placed), although there are some differences with the one used in the experiment. In the Samsung Gear version people do not use a controller to move through the IVE, but instead look at arrows to move in a certain direction. Also, the environment is less immersive than the one of the ORDK. But it might nevertheless be useful to get an impression: https://www.youtube.com/watch?v=digXSxrrjV0

7. In the hardware section, the paper says the software was implemented on one

desktop machine and one laptop. But what happened during the study? Some people used the laptop and some other the desktop?

Response to Reviewer comment No. 7:

Yes, we had one laptop in one room, and the desktop in the other room. Participants were randomly assigned to one of the rooms. We changed the text so that it is clear that we used one laptop and one desktop device during the study (p.22).

8. The measurement section is not always clear too. The first measurement is said to “include current prevention behaviors that are beyond the score (sic) of this paper.” What does this mean? Were people asked about their fire prevention behaviors?

Then the paper says “The prevention behaviors that were measured in the first and third measurement are not analyzed in this paper”. One can try to guess that this was done to assess if there was a behavior change (at least a self-reported ones). If this is the case, it’s strange that the paper omits to analyze it, since the main difference from other research in the literature concerns specifically participant’s behavior.

Response to Reviewer comment No. 8:

In an earlier paper we have studied whether prevention behaviors from various domains (burglary, water damage, fire) can be considered to form a one-dimensional scale (using Rasch analysis). The general conclusion was that these behaviors indeed form a one-dimensional scale. We included all these previously measured prevention behavior items in the first and third measurement, to be able to replicate our previous findings and to see whether the person ordering that the Rasch scale provides, remains consistent under the VR intervention. In this sense, these extra measurements consider the issue of the appropriateness of the Rasch scale measurement, much less the effect of VR. Given that the aim and topic of this measurement was different and our paper already lengthy, we did not include it.

9. To measure the SLOC construct, the paper creates a questionnaire in which some items are supposed to measure internal SLOC while the others external SLOC. However, the section does not report a factor analysis to confirm that the set of created items can actually be divided in the two subscales.

Response to Reviewer comment No. 9:

Based on literature we expected 2 factors. In our SEM we found that the two underlying factors had high correlation, which suggests that we cannot really divide SLOC in two underlying subscales. To confirm this finding we ran a EFA that does indeed shows the ISLOC and ESLOC items load on to different factors, but with many items showing cross loadings and high correlations (>.5 between the two factors). We can supply the results of this analysis, if necessary. NB Other papers differ in whether they treat SLOC as one or two scales: in this sense, finding that both sub-dimensions cannot be distinguished is not rare.

10. One of the items (“How much panic do you think there will be in case of a grease fire?”) that were used to measure severity is different from the usual items that measure that construct, and the choice should be explained.

Response to Reviewer comment No. 10:

De Hoog, Stroebe and de Wit (2008) measured severity with three items, namely how severe, harmful, and serious respondents perceived the health consequences of hypoglycemia. We copied the harmful and severity items. However, the term “serious” translated into Dutch is however also severe (in Dutch “ernst”) or can be translated as “taking something seriously” (in Dutch “serieus”). The latter explanation is however not very relevant for a grease fire. With a health issue such as hypoglycemia someone can take the symptoms and consequences seriously or not. However, it is unlikely to not take a grease fire seriously: anyone will agree that some kind of action needs to be taken. We changed the item to panic this this also reflects the severeness of a fire situation. We now added this explanation in the paper (p. 24).

11. One of the items (“People should be rewarded by their insurance company if they take

preventive measures to prevent or control fire”) that were used to measure SLOC does not actually concern locus of control orientation.

Response to Reviewer comment No. 11:

This item was based on two items of the Aviation Safety Locus of Control Scale (Hunter, 2002), namely:

• Pilots should lose their license if they periodically neglect to use safety devices (for example, seat belts, checklists, etc.) that are required by regulation. (ISLOC)

• Pilots should be fined if they have an accident or incident while “horsing around”. (ISLOC)

The items of the Aviation Safety Locus of Control Scale are negatively formulated: lose license or be fined if they do not follow safety procedures, and our item is positively formulated: people get rewarded if they take safety precautions, but otherwise this is a direct translation of this item.

12. Unlike the first two questionnaires, the third questionnaire was administered on-line. The paper does not provide details about actions taken to prevent issues of on-line surveys, for example: were the answering times measured and taken care of ?

Response to Reviewer comment No. 12:

Whereas in our case we could expect answers to be somewhat more reliable than in an unsolicited survey (people knew beforehand the study also included a third on-line), we tried to take into account the potential unreliability of surveys as much as possible. The survey tool we used (www.mwm2.nl) measures the survey completion time. This allows to exclude those with very small completion times. For longer times matters are less straightforward. When someone does not close the browser, the time will keep on running, so the average or “upper” completion times need not be an indication of unreliability. In the Table below we included the registered completion times per survey. Although the second and the third survey do not exactly include the same items, they are comparable in length (42 resp. 47 items). The minimum completion time in the second survey was 220 seconds and in the third survey 223 seconds, and also the percentage in the lowest time category is similar. This gave us no reason to believe that participants hastily completed the third online survey.

First survey Second survey Third survey

0-299 60.3%* 9.9%** 8.8%***

300-599 39.7% 74.8% 51.9%

600-899 0 14.9% 20.8%

>900 0 .4% 18.5%

N 242 242 216

* The minimum completion time was 157 seconds

** The minimum completion time was 220 seconds

***The minimum completion time was 223 seconds

Besides this, we checked the data for inconsistencies and for instance for multivariate outliers with the BACON algorithm using Stata 14 and did not find any (p. 26).

13. The “descriptive statistics” says that the effect of the IVE is smaller than expected, but provides no effect sizes in the statistics.

Response to Reviewer comment No. 13:

We do not really understand this point and feel we do treat effect size in the text. We indicated that the effect is smaller than we assumed when calculating our sample size: we expected a 20 percentage point difference (assuming 20% for the INFO group vs 40% for the IVE group) (as was mentioned on p. 18). In the experiment we found a difference of 8.6 percentage points (39.8% for the INFO group vs 48.4% for the IVE group) (p.27). We now include the expected difference in the descriptive statistics paragraph (p.27).

In the case the reviewer refers to a lack of mention of effect size measures such as Cohen’s: these are usually provided when testing mean differences between continuous variables. In this case, Cohen’s d can be used in order to overcome the fact that continuous variable distributions can differ in both location and scale, so that the same difference in means could be either large or small, depending on the variation. But probabilities do not have this issue: the difference between the probabilities in the groups is the effect size. We can nevertheless calculate Cohen’s d here and it equals -.172. We did not change the main text, but can add this as desired by the reviewer.

14. In Table 1, the translations from Dutch to English is not always good, in one case “Fire in a home usually exists by the people themselves” the reader cannot be sure of what that is supposed to mean.

Response to Reviewer comment No. 14:

We agree with the reviewer and rewrote some of the items (p.30, 31).

15. The percentage of participants who felt nauseous during the VR experience is surprisingly high and needs discussion. It’s an indication that some VR design guidelines were likely violated by the implemented IVE. This can limit significantly the effectiveness of the IVE for the goal intended in the paper.

Response to Reviewer comment No. 15:

Cyber sickness is a common phenomenon with IVE’s and similar to regular motion sickness, and many individuals suffer from this (cf. Weech, Kenny & Barnett-Cowan, 2019; McCauley & Sharkey, 1992 although both do not give exact estimates). Causal factors can be related to the design of the IVE such as the visual display characteristics (e.g. frame rate), but also to IVE’s in general such as the sensory mismatch between the observed world and the virtual world, or related to the persons such as gender and gameplay experience.

The specifications of the Oculus Rift DK2 Head-Mounted Display (HMD) were:

Resolution: 960 x 1080 per eye

Field of view: 100 degrees diagonal

Latency: 40 ms (estimation of the developer)

Display frame rate: 60 FPS (estimation of the developer)

These specifications are consistent with other VR research and do not point to a violation in the VR design.

As mentioned on p.33 we have tested all the relationships with SEM also on the sample minus the participants who stated they became nauseous during the IVE experience (n = 215). Results showed that all estimated relationships remained stable, indicating that the level of nausea did not affect the results. To give an indication: when performing a chi-square test on the dataset without the nauseous participants, results are similar: in the INFO group 39.8% chose the fire blanket compared to 48.5% (χ2 (1, N = 215) = 1.609, p = .205).

To conclude, although the percentage of nauseous people perhaps might appear high, cybersickness is a very common phenomena (especially when any kind of discomfort can be evaluated as such) and the specifications of the IVE give no reason to doubt violation of the guidelines. Furthermore, the analyses give no argumentation for potential different effects when nausea is avoided.

Weech, S., Kenny, S. & Barnett-Cowan, M., (2019). Presence and Cybersickness in Virtual Reality Are Negatively Related: A Review. Frontier Psychology, 10: 158.

McCauley, M.E. & Sharkey, T.J. (1992). Cybersickness: Perception Virtual Environments. Presence, 1(3).

16. The result that participants using VR learned less than those using printed information needs much more discussion. From the details and screenshots provided, a simple and likely explanation can be traced back to the design and implementation of the IVE, which seems too primitive compared with the state of the art in the VR literature. The paper should discuss the differences in interactivity, realism, and pedagogical methods between the IVE used and the more sophisticated IVEs available in the literature.

Moreover, since the IVE can influence psychological determinants both indirectly (through level of knowledge) and directly, the limitations in the IVE can have had consequences that reach furthest than knowledge only, and weaken the conclusions of the paper. This should be discussed in a critical way.

Care must also be taken in the conclusions section which currently seems to assume that the results hold for IVEs in general, while the limitations in the studied IVE make it difficult to support generalization to more complex IVEs used in learning and training.

Response to Reviewer comment No. 16:

We agree with the reviewer that there are more technically advanced IVE’s, also ones in which people can really move (HTV Vive) or that trigger multiple senses (e.g. heat, wind, fragrance). The IVE was developed in 2014/ 2015 and the study was conducted in 2015, and since then more advanced IVE’s have obviously been available. At the time of the experiment, most of our participants never experienced an IVE (with a head mounted display) and were actually quite impressed by the experience. The mean score for the level of realism of the experience was 3.82 and 3.43 (on a 5 point scale) for the severity of the virtual fire.

The literature that we are referring to used IVE’s of a similar level of interactivity and realism (to our opinion): Chittaro (2012; 2014; 2015; 2016), Zaalberg and Midden (2013), Ahn, Bailenson and Park (2004). Given that previous studies, with IVE’s that are equally or even less advanced (e.g. Chittaro, 2010), have delivered positive results before, and given that most of our subjects had not been introduced to VR before, we had no reason to assume the necessity of a much more elaborate IVE.

However, we do agree with the reviewer that the way in which the knowledge was integrated in the IVE was not optimal, as we had mentioned on p. 40: “An unexpected result was that people in the INFO condition scored higher on knowledge. This could possibly be explained by the fact that the information in the IVE was presented as text in the head-mounted display, which may not have been so easy to read and was perhaps also not compatible enough with the rest of the virtual experience.”

And in line with this remark, on p.40: “One could imagine that the effect of the IVE intervention could be improved by changing the way in which information is put forward. For instance, improved hardware and software might make it easier to produce a static text that is easier to read than the one in our IVE was.”

We integrated a suggestion of how to better integrate knowledge in an IVE as done by Chittaro (2015; 2018). We integrated this on p. 40.

Of course, it is certainly possible that a more technically advanced IVE that reflects a fire even more realistically by better graphics, heat, real smoke etc produces more anxiety and will have larger effects.

If we understand the reviewer correctly, he or she suggests that knowledge might influence vulnerability and severity etc as well, which might affect the conclusions of the analysis. This is, however, not a relation that we have seen in the literature before, nor had we hypothesized it before (although one might be able to come up with an argument along those lines), so we do not feel comfortable adding arrows to the theoretical model as we have it. In addition, mod indices in our SEM models did not suggest such arrows. We can and now do nevertheless address this issue in the discussion (p. 41).

Reviewer #2: In the present manuscript, the authors provide an interesting approach towards fire prevention by using an immersive virtual environment procedure to teach people how to react in the case of a grease fire. The inclusion of actual prevention behavior, together with self-report measures and the use of an experimental setting with randomization and control condition gives this article a promising potential. However, there are some writing and data-analytic issues that, in my opinion, need to be addressed before a recommendation for publication. In my review, I will focus mostly in consistency, writing form and methods which is my area of expertise.

1. The statistical analyses (SEM) done in this article do not take full advantage of the experimental setting created by the authors, given that the results are not controlled by baseline levels (as in a group x time interaction model), which is a problem the authors should address in detail. It is a common practice to measure the constructs in a pre-post fashion, however the authors did not do it to avoid priming. However, the trade-off is that there is no baseline control anymore, turning the analyses correlational, which should be addressed by the authors in more depth.

Response to Reviewer comment No. 1:

We agree that the trade-off between having a baseline and making sure there is no priming is of key importance and in our submission we have argued explicitly about it (p. 44). However, the reviewer’s point that this “turn[s] the analyses correlational” we do not see. Indeed, the analyses are correlational, but they would have been correlational too if we had included a pre-intervention measurement. Both with and without a pre-intervention measurement, the study remains experimental in the sense that the intervention was randomly allocated across participants. Moreover, the extra measurement comes at the expense of additional measurement error, so even the gain in precision that a within-subject manipulation might give, is not guaranteed (see http://datacolada.org/39). We’re not sure whether we should address this point in more detail in the text than we have, but we have reworded the text to make this point clearer and can further elaborate on these issues if wanted, unless the reviewer is suggesting something else here that we apparently miss.

2. On the other hand, I find the data analysis and results section rather confusing. This is, in my opinion, mostly because it is very data dense and contains a lot of modelling, which always creates a challenge in reporting. One consequence of this, is the omission of important information to evaluate the quality of the data analyses. This is especially relevant for the CFA and SEM parts, where they show some inconsistencies that I’m listing below.

pp 6 (and others). When talking about psychological determinants, please be more specific. Psychological determinants of prevention behavior?

Response to Reviewer comment No. 2:

We included a sentence on p. 3 to make this clear for the remainder of the paper.

3. pp 7: "The underlying arguments are equally appropriate for the area of fire prevention behavior" It is important to give arguments when presenting such a statement, why?

Response to Reviewer comment No. 3:

We now include this argumentation on p. 7.

4. Pp 7. "Studies usually use" is not the best argument for selection of variables: I strongly suggest the authors to describe a sound logical and empirical basis for their proposal. This is done later; however, it can be briefly explicated here.

Response to Reviewer comment No. 4:

We cannot find a sentence with “studies usually use” in our manuscript. We think the reviewer referred to this sentence:

“IVE studies often also consider knowledge about the topic and locus of control as variables that may be influenced by IVE and may themselves influence subsequent prevention behavior”.

In any case, we agree with the reviewer and have adjusted the text (p. 7).

5. Pp.7 "We feel it is appropriate or at least worthwhile to study psychological determinants and the target behaviors simultaneously" Why? It is, indeed, important, but this should be elaborated.

Response to Reviewer comment No. 5:

We now elaborate on this point (p. 8).

6. pp 8. "Can in fact be studied empirically relatively easily" How?

Response to Reviewer comment No. 6:

We were referring to studies in the aircraft evacuation domain (e.g. Chittaro 2012; 2014; 2015; 2016) in which it is not that easy (to say the least) to measure whether people follow the correct safety procedures during an accident. In the case of fire prevention measures, measuring the real behavior that we consider is way easier, since we can just measure the purchase of the fire blanket. On the other hand, measuring if people will follow the correct safety procedures during a fire, indeed faces the same problems as in the aircraft evacuation domain. We decided the remove the sentence (p. 8).

7. pp11. Please check repeated citation (29,29).

Response to Reviewer comment No. 7:

Now corrected (p. 11).

8. pp11. Please provide references for the SLOC bi-dimentionality, if they were provided before, I suggest moving the definition of the construct together with the citations above.

Response to Reviewer comment No. 8:

Now integrated (p. 10).

9. pp 22. Why are only knowledge, vulnerability and severity5 included in the third measurement? This exclusion should be supported.

Response to Reviewer comment No. 9:

We now explain this exclusion (p. 23).

10. Measures: Please report basic psychometric properties of the instruments here (cronbach's alpha for internal consistency, for example, and factor structure of original instruments).

Response to Reviewer comment No. 10:

The psychometric properties (Cronbach’s alpha, factor loadings and AVE) for the eventually chosen items/ scales are given in Table 1 on p. 30. We also supplied the data for the process of removing items that do not fit well, but based on AVE (following the procedure as in Knijnenburg &Willemsen, 2015), not based on factor loadings or Cronbach’s alpha. We added an Appendix to this document (“Appendix for reviewers”) in which the factor loadings and Cronbach’s Alpha’s are presented for the original scales (Table 1). We can add this Table to the paper if considered useful according to the reviewer.

Knijnenburg, B.P., Willemsen, M.C. (2015). Evaluating recommender systems with user experiments. In: Ricci, F., Rokach, L., Shapira, B. (eds.) Recommender Systems Handbook, pp. 309–352. Springer, New York.

11. pp. 27. Please specify the software version used for SEM analyses.

Response to Reviewer comment No. 11:

Now included (p. 28).

12. Data analyses and report:

Please describe which sampling wave was used for the analyses (I guess it’s sample 2, however I cannot find it in the manuscript).

Response to Reviewer comment No. 12:

It is indeed sample 2. This was mentioned on p. 23:

“The hypothesized model (Fig 1) is based on differences between the IVE and the INFO group at the time of the second measurement.”

13. Measurement models: I find it problematic to test all measurement models together, given that this would further obscure local sources of bad fit. Is it just one instrument the one creating the bad fit? Are all measurement models wrong? I would suggest test this separately. On the other hand, the procedure of co-variation between items within an instrument should be done in this step.

Response to Reviewer comment No. 13:

Opinions differ on whether estimating the model in its entirety or estimating the parts separately is the better approach. We followed the procedure to do CFA on multiple factors in one single estimation as suggested Brown (2006) and performed by Knijnenburg and Willemsen (2015).

Brown, T. A. (2016). Methodology in the social sciences. Confirmatory factor analysis for applied research. New York, NY, US: The Guilford Press.

Knijnenburg, B.P., Willemsen, M.C. (2015). Evaluating recommender systems with user experiments. In: Ricci, F., Rokach, L., Shapira, B. (eds.) Recommender Systems Handbook, pp. 309–352. Springer, New York.

Nevertheless, we have re-run our models with separate measurement models for each factor individually, and for the accompanying modified models. See the enclosed Appendix, Table 2-5 for the model fit statistics and Table 6-7 for the R-squared estimates and AVE’s. It should be noted that when performing CFA on individual factors with only 3 indicators this shows a perfect fit because these are “just-identified” (df=0) (see Table 2-3). When performing CFA’s on the individual factors similar results apply compared to testing the measurement models all together, and the same items should be removed to improve AVE. The Appendix is quite lengthy and offers little extra value to the paper, but we can of course include it if wanted.

14. Structural Equation Models: Modifications were made to the models, which is a step towards a good fit. It is important to see if the modifications suggested by the modification indexes are reasonable under theoretical grounds. This is done for the co-variation between variables; however, it should also be made for the covariation between items. It is usually accepted to co-vary items within a scale if they share method or another known source of shared variance that should not be captured by the latent factor (please see previous comment, however, regarding the location of this procedure).

Response to Reviewer comment No. 14:

We did not have any theoretical grounds for adding the covariation between items: correlation between any of the items in principle is not unreasonable though. The main reason to add them was to create maximum fit model and see whether this impacts on our results. As mentioned on p. 32: “We only incorporated this step to show how model fit could be improved, but will not elaborate on this step, as adding correlations at the item level is not very common.” We indeed found that adding these correlations does not affect our conclusions. We now left them in, but we can also leave model 3 and 5 out of the paper, as desired the reviewer.

15. It is not clear why do the authors base their report on model 4, if the model with a good fit is model 6. It is important to report the good fitting model. Even though models showed no substantial changes in their point estimates, the non-significant results are part of a bad-fitting model and thus are not directly interpretable (i.e. SLOC). This should be at least stated as a limitation or tested with a different regression technique.

Response to Reviewer comment No. 15:

The models that we present represent different trade-offs between on the one hand theoretical rigor (what did we hypothesize beforehand) and empirical fit (which covariations can or should we take into account to make sure the measurement model fits as best it can). We chose Model 4 for our conclusions as the one that remains closest to our a priori theoretical arguments and nevertheless has a decent fit but, as can be seen in the paper, reporting the results based on Model 6, as the reviewer suggests, is possible too and causes no changes in the statistically significant effects (and of course has better fit statistics).

We do not see why the non-significant results are the necessary consequence of a poorly fitting model (as opposed to showing that there is no true effect), given that Models 4 and 6 lead to essentially the same results, although we could report Model 6 in the text instead (we did not do this now). Perhaps we are misunderstanding the reviewer here, or perhaps the reviewer mixed up some things. The SLOC variable that the reviewer mentions is not in the estimated model at all.

16. I find it important to explicitly say, on the SEM models, how were the dichotomous outcomes treated, and interpret them accordingly. Where they estimated using a probit or logistic regression? (In Mplus this is different for different estimation methods)? Please clarify this on the manuscript.

Response to Reviewer comment No. 16:

We used Weighted Least Squares, which goes with probit regression in M-Plus (maximum likelihood would have used logistic regression). We added this information on p. 28.

17. Please also describe which method did the authors use to compute the mediation analyses (multiplicative?) and provide the estimates in the tables.

Response to Reviewer comment No. 17:

Indirect effects in Mplus are indeed calculated by multiplying the direct effects. The results of this, including the bootstrapped results, are in running text (p. 37). Do you want us to provide the results in the text and in a separate table?

18. Pp 36. Please specify which bootstrap method was used. Bias corrected, percentile? They have shown different trade-offs regarding type 1 errors.

Response to Reviewer comment No. 18:

We used the bootstrap percentile method. We now mention this in the text on p 37.

19. Pp 36. A direct effect of IVE on fire blanket is reported, however later it is exposed that no significant effect of IVE on behavior was found. Please clarify this.

I think here the problem may arise because the total effect is non-significant while the direct effect is significant when the mediator is partialized. By checking the estimates, it is possible to see, as the authors indicated, that indirect and direct effects have opposite signs. This, together with the fact that the direct effect is bigger than the total effect is an indication of an inconsistent mediation, where the mediator acts as a suppressor variable in the relationship between IVE and blanket. Given that the mediator is part of the model, this may be interpreted as a significant direct effect when the mediator is taken into account, and should be discussed accordingy.

Response to Reviewer comment No. 19:

The interpretation by the reviewer is indeed correct. Perhaps the confusion arose because one could argue that there is a positive direct effect when controlling for the mediator. Instead we describe our findings in terms of the total effect (none) and two separate significant effects: a positive direct effect and a negative indirect effect through the mediator. We now clarified that the total effect we found is non-significant at p. 37.

20. Pp 36. A nonsignificant effect is presented as "marginally significant". I suggest avoiding using this language on the manuscript because they can be misleading.

Response to Reviewer comment No. 20:

Now adjusted on p. 37.

21. Pp 37. Please provide assumption checks for the repeated measures anova, and report results in a corresponding table.

Response to Reviewer comment No. 21:

Good point. We see now that the assumption for normally distributed residuals was violated (checked with Shapiro-Wilk test). We therefore performed the analyses again, using a multilevel regression analysis instead, with a robust estimator correcting for the non-normally distributed residuals. Results show some changes in significances, as reported on p. 38, 39. The conclusions about the interaction effects remain the same.

We provide the changed results in the text. We did not really get which results the reviewer wanted us to report in a table (what is left to report?), but are willing to do so, if necessary.

22. Discussion:

The authors propose here that the relationship between IVE and self-efficacy is partly mediated by vulnerability, however in their model (figure 3) they estimate a co-variation and not an indirect effect properly. Please clarify this point, and in the case there is an indirect effect, please report it accordingly. If no mediation effect was computed, this should be not discussed as such.

Response to Reviewer comment No. 22:

We think this might be a misunderstanding. We propose the relationship between IVE and self-efficacy is partly mediated by severity, not vulnerability, and we estimated it as such. In figure 3 we have single headed arrows for relationships between concepts and double headed arrows for correlations between concepts. These correlations between concepts arise from the mod-indices.

23. Please provide a substantial interpretation of the significant mediation effect of IVE to blanket towards knowledge, so the reader can get an idea of the mechanism involved.

Response to Reviewer comment No. 23:

We integrated an interpretation of the mediation effect of knowledge (as a mediator between IVE - fire blanket), and we added an explanation for the missing mediation effect (knowledge as a mediator between IVE- flyers, as was hypothesized) (p.41, 42).

24. The authors discuss the nonsignificant interaction effect. In this case, it is advisable to avoid discussing nonsignificant results as “marginally significant”. The authors describe this as a “fully” mediated effect, which may not be the case given what I exposed before.

Response to Reviewer comment No. 24:

We agree. We maintained the paragraph that discusses the direct and indirect results, but changed the wording in both this paragraph and the discussion section in a way that reflects the suggestion by the reviewer (p. 37, 41, 42).

25. The authors discuss the importance of causal claims; however, this is beyond the scope of the present article (I describe this at the beginning). Same for the influence of IVE on the psychological domains, it is important to control for previous values.

Response to Reviewer comment No. 25:

Please see our comment in the beginning, where this point comes up before: having an earlier measurement (“previous values”) would have been possible but we deliberately did not do this as we outlined above. Nevertheless, the design remains experimental: participants were randomly divided over 2 conditions.

General comments:

26. The authors usually justify their decisions with statements such as "We feel it is appropriate or at least worthwhile" (pp 7), "we felt it was relevant to compare the IVE experience with an information sheet". Although their decisions seem reasonable from a methodological point of view, they need to be justified based on methodological or theoretical grounds (i.e. benefit of simultaneous measurement of self-report and behavior, benefit of a control condition versus no control condition).

Response to Reviewer comment No. 26:

We added justifications in the text (p. 8, 9).

27. Please state the difference between serious and commercial games.

Response to Reviewer comment No. 27:

We think the reviewer is referring to this sentence:

“Kato, Cole, Bradlyn, and Pollock (2008) have shown a positive effect of a 3-month serious game on knowledge about cancer treatment, compared to a control group who received a commercial game.”

In this case the commercial game was the video game Indiana Jones and the Emperor’s Tomb.

We integrated this, and also provided a definition of a serious game (p. 9).

I hope the authors find this review useful for their research!

We thank the reviewers for their time, effort, and useful comments.

Best regards,

________________________________________

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: Yes: Cristóbal Hernández C.

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files to be viewed.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org. Please note that Supporting Information files do not need this step.

Attachment

Submitted filename: Response to Reviewers.docx

Decision Letter 1

Geilson Lima Santana

23 Dec 2019

PONE-D-19-17627R1

Playing with fire. Understanding how experiencing a fire in an immersive virtual environment affects prevention behavior.

PLOS ONE

Dear Ms. Jansen,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

We would appreciate receiving your revised manuscript by Feb 06 2020 11:59PM. When you are ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter.

To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). This letter should be uploaded as separate file and labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. This file should be uploaded as separate file and labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. This file should be uploaded as separate file and labeled 'Manuscript'.

Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out.

We look forward to receiving your revised manuscript.

Kind regards,

Geilson Lima Santana, M.D., Ph.D.

Academic Editor

PLOS ONE

Additional Editor Comments (if provided):

Thank you for incorporating reviewers's advices.

I believe it is important to include the appendix to reviewers. This would help interested readers have a deeper understanding of the methods and results found.

Please, in your manuscript, you need to include not only page numbers, but also line numbers. Use continuous line numbers (do not restart the numbering on each page).

[Note: HTML markup is below. Please do not edit.]

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files to be viewed.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2020 Mar 6;15(3):e0229197. doi: 10.1371/journal.pone.0229197.r004

Author response to Decision Letter 1


27 Jan 2020

1.) We added the appendix to the manuscript, labeled as S4 File (as agreed to by e-mail by Sarah Mills, on Jan 24th 2020).

2.) We added line numbers in the manuscript.

Attachment

Submitted filename: Response to Reviewers_track changes.docx

Decision Letter 2

Geilson Lima Santana

3 Feb 2020

Playing with fire. Understanding how experiencing a fire in an immersive virtual environment affects prevention behavior.

PONE-D-19-17627R2

Dear Dr. Jansen,

We are pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it complies with all outstanding technical requirements.

Within one week, you will receive an e-mail containing information on the amendments required prior to publication. When all required modifications have been addressed, you will receive a formal acceptance letter and your manuscript will proceed to our production department and be scheduled for publication.

Shortly after the formal acceptance letter is sent, an invoice for payment will follow. To ensure an efficient production and billing process, please log into Editorial Manager at https://www.editorialmanager.com/pone/, click the "Update My Information" link at the top of the page, and update your user information. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to enable them to help maximize its impact. If they will be preparing press materials for this manuscript, you must inform our press team as soon as possible and no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

With kind regards,

Geilson Lima Santana, M.D., Ph.D.

Academic Editor

PLOS ONE

Acceptance letter

Geilson Lima Santana

21 Feb 2020

PONE-D-19-17627R2

Playing with fire. Understanding how experiencing a fire in an immersive virtual environment affects prevention behavior.

Dear Dr. Jansen:

I am pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please notify them about your upcoming paper at this point, to enable them to help maximize its impact. If they will be preparing press materials for this manuscript, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

For any other questions or concerns, please email plosone@plos.org.

Thank you for submitting your work to PLOS ONE.

With kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Geilson Lima Santana

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 File. Text for INFO group (translated from Dutch).

    (DOCX)

    S2 File. Text in IVE (translated from Dutch).

    (DOCX)

    S3 File. Knowledge scale (translated from Dutch).

    (DOCX)

    S4 File. Additional results of Confirmatory Factor Analysis.

    (DOCX)

    S1 Table. Actions people took in the IVE fire game, during the first and second game play.

    (DOCX)

    Attachment

    Submitted filename: Response to Reviewers.docx

    Attachment

    Submitted filename: Response to Reviewers_track changes.docx

    Data Availability Statement

    All data files are available from the Open Science Framework database. Jansen, P.C.P., Snijders, C.C.P. & Willemsen, M.C. (2020, February 14). Dataset: effect of IVE on prevention behavior. Retrieved from osf.io/kwq45 DOI 10.17605/OSF.IO/KWQ45.


    Articles from PLoS ONE are provided here courtesy of PLOS

    RESOURCES