Skip to main content
PLOS One logoLink to PLOS One
. 2023 Jan 30;18(1):e0267983. doi: 10.1371/journal.pone.0267983

The impact of visually simulated self-motion on predicting object motion–A registered report protocol

Björn Jörges 1,*, Laurence R Harris 1
Editor: Michael J Proulx2
PMCID: PMC9886253  PMID: 36716328

Abstract

To interact successfully with moving objects in our environment we need to be able to predict their behavior. Predicting the position of a moving object requires an estimate of its velocity. When flow parsing during self-motion is incomplete–that is, when some of the retinal motion created by self-motion is incorrectly attributed to object motion–object velocity estimates become biased. Further, the process of flow parsing should add noise and lead to object velocity judgements being more variable during self-motion. Biases and lowered precision in velocity estimation should then translate to biases and lowered precision in motion extrapolation. We investigate this relationship between self-motion, velocity estimation and motion extrapolation with two tasks performed in a realistic virtual reality (VR) environment: first, participants are shown a ball moving laterally which disappears after a certain time. They then indicate by button press when they think the ball would have hit a target rectangle positioned in the environment. While the ball is visible, participants sometimes experience simultaneous visual lateral self-motion in either the same or in the opposite direction of the ball. The second task is a two-interval forced choice task in which participants judge which of two motions is faster: in one interval they see the same ball they observed in the first task while in the other they see a ball cloud whose speed is controlled by a PEST staircase. While observing the single ball, they are again moved visually either in the same or opposite direction as the ball or they remain static. We expect participants to overestimate the speed of a ball that moves opposite to their simulated self-motion (speed estimation task), which should then lead them to underestimate the time it takes the ball to reach the target rectangle (prediction task). Seeing the ball during visually simulated self-motion should increase variability in both tasks. We expect to find performance in both tasks to be correlated, both in accuracy and precision.

Introduction

We are constantly immersed in complex, dynamic environments that require us to interact with moving objects, for example when passing, setting, or hitting a volleyball, or when deciding whether we can make a safe left turn before another car reaches the intersection. In such situations, it can often be important to predict how objects in our environment will behave over the next few moments. Predictions allow us, for example, to time our actions accurately despite neural delays [13] in perceiving moving objects and issuing and executing motor commands [4, 5]. Delays of between 100ms and 400ms between visual stimulation and motor response are generally assumed [6]. Without predicting or anticipating motion, we would thus always be acting on outdated positional information and be doomed to fail when attempting to intercept moving objects, particularly when they are relatively fast. Further, and perhaps more obviously, predictions are also important when moving objects are occluded during parts of their trajectory [79], or when the observer has to avert their eyes or turn away from the target.

When an observer is moving while attempting to interact with moving objects in their environment, further difficulties arise. Even in the simplest case, when the observer has a good view of the target and predictions are only necessary to compensate for neural delays, the visual system needs to separate retinal motion created by observer motion from retinal motion due to object motion in order to judge an object’s trajectory accurately. A prominent hypothesis on how this is achieved is the Flow Parsing hypothesis [1016]. This hypothesis states that to solve this problem humans parse optic flow information and subtract this visual stimulation attributed to self-motion from global retinal motion. The remaining retinal motion can then be scaled with an estimate of the distance between the observer and the moving object [17] to obtain the world-relative velocity of the object. While this process was originally posited as a purely visual phenomenon [18, 19], more recent studies have shed some light on the multisensory nature of this process: Dokka and her colleagues [20] found, for example, that compensation for self-motion in speed judgements was more complete when both visual and vestibular information indicated self-motion than when either visual or vestibular cues indicated that the observer was at rest. They further found precision to be lower in the presence of self-motion, which they attributed to self-motion information being noisier than optic flow information. Subtracting noisy self-motion information [21] from less noisy optic flow [20] would then make the estimate of object motion noisier in a moving observer than in a static observer. It is important to note that such a flow parsing mechanism should be active even while the participant at rest. The assumption here is that the noise added during flow parsing is proportional to the self-motion speed (in a Weber’s Law-like fashion), that is, when the observer is not moving at all the added noise is minimal, whereas higher self-motion speeds lead to more noise.

Several studies have shown that flow parsing is often incomplete [9, 2227]. A candidate explanation for this incompleteness is an inaccurate estimation of self-motion: typically, studies have used stimuli where only some cues indicated that the observer was moving (usually visual and/or vestibular cues), while efferent copy cues indicated that their bodies were at rest. If all self-motion cues are integrated after weighting them (e.g., according to their relative reliabilities [21, 28, 29]), this would then lead to an underestimation of self-motion, which is consistent with the biases found in the studies cited above. Further, while we did not find evidence for an effect of self-motion on precision in a recent study [30], we believe it is likely due to the fact that the effect was noisier than anticipated, resulting in a lack of statistical power.

Fig 1 shows a simple schematic of the processes we assume to be at play when predicting object motion during self-motion: the organism first estimates its own motion in the environment from the various sources of information available to it. Based on this self-motion estimate, the organism then makes a prediction about the retinal motion that this self-motion would be expected to create. The predicted retinal motion is then subtracted from the total observed retinal motion and any remaining motion is attributed to the object, a process we call “multisensory flow parsing” to distinguish it from the purely visual conceptualization of flow parsing brought forward by authors like Wertheim [15, 16] or, more recently, Dupin and Wexler [10]. This step adds noise to the velocity estimate because the self-motion estimate is noisier than the retinal input [20]. The further trajectory of the object would then be extrapolated based on this estimate of its velocity.

Fig 1. Schematic of the processes at play when predicting motion during self-motion.

Fig 1

First, an estimate of the self-motion velocity is obtained by integrating the available cues from different sensory sources. This velocity estimate is used to predict the retinal motion that would be caused by self-motion. Finally, an estimate of the physical object velocity is obtained by subtracting the predicted retinal motion due to self-motion from the global retinal motion and the further trajectory of the object is extrapolated based on this estimate.

This schematic provides an overview over the general mechanism that might be at play when predicting future motion of an object observed during self-motion. Depending on the motion profiles of observer and object further complications may arise. For example, retinal speeds depend not only on the physical speeds and directions of the observer and object but also on the distance between them and can therefore change systematically even when observer and object move at constant physical speed without changing direction. To obtain a veridical representation of the physical velocity of the object, the observer thus has to perform additional computations, including estimating their distance to the object, the direction of the object in relation to their own direction of motion, and the necessary transformations to obtain the physical object speed from these values [17].

Some studies suggest that biases incurred while estimating motion, e.g., due to the Aubert-Fleischl effect which lowers the perceived speed of a target when an observer tracks it with their gaze [31], or due to low contrast in the stimulus [32], might transfer to biases in motion extrapolation based on these speed estimates. It seems straight-forward that any biases and precision differences observed in the perception of speed would correlate perfectly with errors and precision differences in time-to-contact judgements. However, there are two complications: first, participants might integrate biased and less precise speed information obtained during self-motion with prior information they have formed in response to previous exposure to the stimulus. They might thus extrapolate motion biased on a combination of prior information and (biased and more variable) online information. Further, it has been reported that under certain circumstances perceptual judgements and action-related tasks can be based on separate cues (see, e.g., [3335]). While it remains an appealing hypothesis, it should thus not be assumed that biases and variability differences in time-to-contact judgements reflect only biases and variability differences in speed estimation. Studying to what extent biases and variability differences in online information acquired while viewing a target influence the way we extrapolate its further motion will help us better understand the predictive mechanisms at play not only when the target is occluded or the observer averts their gaze from the target but also when timing interceptive actions accurately despite unavoidable neural delays [36]. In the present study, we therefore investigate how biases in speed estimation elicited by visual self-motion impact the prediction of object motion.

More specifically, we will test three interconnected hypotheses:

  • Predictions of where a moving object will be in the future will be biased (Hypothesis 1a) and more variable (Hypothesis 1b) in the presence of visually simulated self-motion

  • Object speed judgements will be biased (Hypothesis 2a, which constitutes a replication of an earlier result of Jörges & Harris, 2021) and more variable (Hypothesis 2b, a more highly powered follow-up to a hypothesis for which Jörges & Harris, 2021 did not find significant support) in the presence of visually simulated self-motion

  • The effect of visually simulated self-motion on motion extrapolation can be predicted from its effect on speed estimation, both in terms of bias (Hypothesis 3a) and its variability (Hypothesis 3b)

Methods

Apparatus

We programmed all stimuli in Unity 2020.3.30f1. Given the on-going COVID-19 pandemic, some participants who are owners of head-mounted VR devices (HMDs) are tested in the safety of their home. Our experiment is compatible with all major modern HMDs. To minimize variability introduced by the use of different HMDs, each program we use to present stimuli limits both the horizontal and the vertical field of view to 80°. We don’t expect there to be relevant differences in frame rate, as Unity caps the frame rate at 60 Hz. Our programs are, further, much less demanding in processing power than any VR application remote participants are likely to run; that is, frame rate dropping below 60 Hz should occur almost never. If it is safely possible and an ethics approval is granted, we might also test some participants in person in our laboratory. For these participants, we will use an VIVE Pro Eye.

Participants and recruitment

We recruit participants with HMDs in their possession online for them to perform the experiment in their homes. Recruitment occurs through social media (such as Twitter, Facebook, and Reddit). Some remote participants might also be recruited through the professional recruitment service XpertVR. Since recruiting participants with VR equipment at home is not trivial, we might also rely on York University participant pools to recruit participants for in-person testing. In this case, all applicable guidelines for safe face-to-face testing are fulfilled and exceeded. Participants receive a monetary compensation of 45 CAD for participation in the experiment; participants recruited through the York University participant pools may receive course credit instead of a monetary compensation. All participants are screened for stereo-blindness with a custom Unity program (downloadable on GitHub: https://github.com/b-jorges/Stereotest) in which participants have to distinguish the relative depth of two objects that are matched in retinal size. Participants are only included if they answer correctly on 16 out of 20 trials. The simulated disparity is 200 arcsec. While this allows only for a coarse assessment of the participants’ stereovision, our experiment is not critically dependent on a high stereoacuity. We test 20 men and 20 women (see Power Analysis). The experiment was approved by the local ethics committee and is conducted in accordance with the Declaration of Helsinki.

Stimulus

Each participant performs two main tasks in an immersive VR environment: a prediction task and a speed estimation task. Every participant completes both tasks, and we counterbalance the order in which they complete them such that 20 participants (10 women and 10 men) start with the prediction task, and 20 participants (10 women and 10 men) start with the speed estimation task. All programs we used to present the stimuli are available on Open Science Foundation (https://osf.io/gakp5/), and the Unity projects can be downloaded on Open Science Foundation as well (https://osf.io/6mz4w/).

For both tasks, we display a circle in the middle of the observer’s field of view that moves with their head rotation in front of the Earth-stationary simulated world. Participants are instructed to keep this circle surrounding the fixation cross. When the center of the circle is with 2.5° (vertically and horizontally) of the fixation cross, it disappears to indicate to the participant that their head is positioned correctly. We further record head position whenever the ball or the ball cloud (see below) are visible. Since recording head position on each frame might slow down the program on older systems, we opt to record the mean rotation (horizontally and vertically) over bins of five frames.

Prediction

In the prediction task (see Fig 2A, and see also this video on OSF (https://osf.io/rkg23/), we first show participants a ball of 0.4 m diameter moving laterally 8m in front of them at one out of three speeds (4, 5, 6 m/s). We have used this range of speeds in our previous study [30], while the size of the ball was diminished slightly in comparison to this study (see description of the speed estimation task for the rationale). The ball can travel to the left or to the right. It appears to the left of the observer when it travels to the right, and on the right of the observer when it travels to the left such that the visible part of the trajectory is centered in front of the observer. At the same time, a target rectangle is presented on the side towards which the ball is travelling. The ball disappears after 0.5s and participants press the space bar on their keyboard in the moment they think the ball would hit the target. In order to curtail biases in speed perception due to the Aubert-Fleischl phenomenon [37], participants are asked to keep their eyes on a fixation cross that is presented straight ahead of them and slightly below the stimulus (see Fig 2A) and moves with the observer when they experience visual self-motion. The target is presented at a distance that depends on the speed of the ball and the occlusion duration, which can be 0.5 s, 0.6 s, or 0.7 s. Speeds and occlusion durations are chosen such that, when the participant keeps their gaze on the fixation cross, the whole trajectory (including the invisible part) unfolds within a field of view of 60°, which is well within the effective field of view of any modern HMD. The distance between the point where the ball disappears (the “point of disappearance”; see Fig 2C) and the target rectangle is given by the following equation:

Distance=DurationOcclusion*SpeedBall [1]
Fig 2.

Fig 2

A. Screenshot from the prediction task while the ball was visible. B. Screenshot from the speed estimation task while the ball cloud was presented. C. Schematic of the prediction task. D. Schematic of the speed estimation task.

While the target is visible, participants experience lateral visual self-motion either in the same direction as the ball or in the opposite direction as the ball, or they remain stationary. The self-motion speed ramps up in a Gaussian fashion over the first 50 ms until it reaches 4 m/s, then remains constant for 400 ms, and finally ramps down again over the last 50 ms before the ball becomes invisible. Overall, the observer moves 1.8 m over 500 ms. Please note that the different motion profiles elicit very different retinal speeds: observer motion in the opposite direction of the ball elicits higher retinal speeds overall than for a static observer or for observer motion in the same direction as the ball. Table 1 displays the mean absolute retinal speeds across the trajectory for all conditions. While our previous results [30] suggest that the role of the retinal speeds for the overall precision in motion estimation is subordinate to other sources of variability, retinal speeds are highly correlated with the expected effect of our self-motion manipulation on variability.

Table 1. Mean absolute retinal speeds across the visible part of the trajectory for each combination of ball speed and observer motion profile.

The script in which we derive these values can be found on GitHub (https://github.com/b-jorges/Predicting-while-Moving/blob/main/Geometry%20Prediction.R).

Ball Speed
4 m/s 5 m/s 6 m/s
Observer Static 0.4°/s 28.4°/s 34.0°/s
Same Direction 22.2°/s 5.9°/s 11.6°/s
Opposite Directions 44.4°/s 50.4°/s 55.8°/s

We further add a range of occlusion durations (0.1s, 0.2s, 0.3s, 0.4s, 0.8s, 0.9s, 1s) while the observer is static to get an estimate of how variability changes in response to different occlusion durations. Overall, participants complete 225 trials (3 ball speeds * 3 self-motion profiles * 3 occlusion durations * 5 repetitions + 3 ball speeds * 6 occlusion durations * 5 repetitions), which takes around 10 minutes.

Participants complete a brief training of 18 trials before starting the main experiment (see this video on OSF: https://osf.io/4js5w/). The ball travels at one of three speeds (2.5, 3.5, 4.5 m/s), going either left or right, and three occlusion durations (0.45, 0.55, 0.65 s). In the training, the ball reappears upon pressing the spacebar in the position it would have been in at that moment. No visual self-motion is simulated in the training. This allows participants to estimate their error (spatially) and helps them familiarize themselves with the task and the environment.

Speed estimation

In the speed estimation task (video on OSF: https://osf.io/xqkgy/), participants are presented with two motion intervals and have to judge which of them is faster. In one interval, they view a ball travelling to the left or to the right. As for the prediction task, this ball can have one of three speeds (4, 5, 6 m/s), and the participant can also experience visual self-motion in the same direction or in the opposite direction or remain static (see Fig 2D). Except for the self-motion intervals, the scene is static just like in the prediction experiment, and object motion occurs relative to the scene. The second motion consists of a ball cloud of 2.5 m width and 1 m height at the same distance to the observer (see Fig 2B). Each ball in this cloud has the same diameter as the main target (0.4m) and balls are generated at one side of the cloud and then move to the other side where they disappear. In our previous study [30], we used smaller balls for the ball cloud, which may have been a factor their judgements of speed being consistently overestimated relative to the single ball. We therefore decided to use the same ball size for both the single ball and the elements of the ball cloud. At any given moment, between 8 and 12 balls are visible. All the balls in the ball cloud move at the same speed and are visible either until they reach the opposite side of the cloud area or until the motion interval ends after 0.5s observation time. The speed of these balls is constant throughout each trial and is governed by PEST staircase rules [38]. For each condition, we employ two pests: one starts 30% above the speed of the single ball from the other motion interval, while the other starts 30% below. The initial step size is 0.6 m/s and each pest terminates either after 37 trials, or when the participant has completed at least 30 trials and the step size drops below 0.03 m/s. We modified the original PEST rules such that the step size is always twice the initial step size (that is, 1.2 m/s) for the first ten trials in order to spread out the values presented to the observer and allow for more robust JND estimates. We limit the range of speeds the ball cloud can take to between one third of the speed of the single ball and three times the speed of the single ball. Participants are asked to maintain fixation on the fixation cross that had the same characteristics as in the prediction task, that is, it is always presented slightly below the stimulus and it moves with the participant as they experience visual self-motion. To keep the visual input identical across both tasks, the target rectangle from the prediction task, while irrelevant for the speed estimation task itself, is present in this task as well. Overall, participants perform between 30 and 37 trials in 18 staircases (two start values, three speeds and three motion profiles) for a total of between 540 and 666 trials.

Before proceeding to the main tasks, participants complete a training session. This training session consists of one PEST of reduced length (between 20 and 27 trials) that starts 30% above the speed of the ball (3 m/s). Participants need to achieve a final step size of below 0.3; otherwise, they are asked to repeat the training. If they fail the training a second time, we exclude them from the analysis. The participants do not experience visually simulated self-motion in this training. This task–including the training–takes about 40 minutes to complete.

Participants can choose to receive the instructions as PDF (can be downloaded from GitHub: https://github.com/b-jorges/Predicting-while-Moving/blob/main/Instructions%20Predicting%20while%20moving.pdf) or watch a video (which can be viewed on YouTube: https://youtu.be/qHTWVyjn0QI and https://youtu.be/JyOZ-duRGmU, respectively).

Modelling the predictions

To obtain specific predictions corresponding to each hypothesis, we built models of the underlying perceptual processes for both the prediction and the speed estimation task. The instantiation of the model for the prediction task can be found here (on GitHub: https://github.com/b-jorges/Predicting-while-Moving/blob/main/Analysis%20Prediction.R), and the instantiation of the speed estimation model can be found here (on GitHub: https://github.com/b-jorges/Predicting-while-Moving/blob/main/Analysis%20Speed%20Estimation.R). The implementation of the model that relates performance in both tasks can be found here (on GitHub: https://github.com/b-jorges/Predicting-while-Moving/blob/main/Predictions%20Correlations.R). While a detailed discussion of these models can be found in S1 Appendix, the most important assumptions are the following. The first assumptions reflect our hypotheses:

  • The speed of the ball is overestimated by 20% of the presented self-motion speed when observer and ball move in opposite directions [30], with a between-participant standard deviation of 30%. In the prediction task, this overestimation of speed should lead to an underestimation of the time it takes for the ball to travel the occluded distance. No biases are assumed for the Same Directions and Observer Static motion profiles. This assumption reflects Hypothesis 2a.

  • While we previously did not find evidence for an impact of visual self-motion on precision [30], we believe that the higher self-motion speeds in this study might enable us to uncover small effects that were not apparent at lower self-motion speeds. The variability in perceived speed is 20% higher for the Opposite Directions motion profile, with a between-participant standard deviation of 30%. No differences in variability are assumed for the Opposite Directions and Static motion profile. This assumption reflects Hypothesis 2b.

  • The same effects of self-motion on accuracy and precision are also at play in the prediction task. This assumption reflects Hypotheses 1a and 1b.

  • Participants display the same biases in perceived object speed in response to self-motion in the opposite direction of the ball in both the speed estimation and the prediction task. Similarly, variability is impacted equally in both tasks. This assumption reflects Hypotheses 3a and 3b.

We further need to make several assumptions about how the participants process the stimulus independently of the presence of visual self-motion:

  • We neglect how exactly participants recover physical speed from angular speed and the perceived depth but to acknowledge the added complexity, we assume a Weber Fraction of 10% for the estimation of the ball speed, which is slightly higher than is generally reported for less complex speed discrimination tasks [39], with a between-participant standard deviation of 1.5%.

  • Prediction task only: We assume that the computations executed by the visual system are approximated accurately with the physical equation for distance from speed and time (d = v*t), such that the extrapolated time (textrapolated) can be estimated from the distance between the point of disappearance and the target (dperceived) and the perceived speed of the ball (vperceived):

textrapolated=dperceived/vperceived (2)
  • Prediction task only: The distance between the point of disappearance of the ball and the target rectangle (that is, the occluded distance) is estimated with a Weber Fraction of 5%, as reported in the literature [40]. We further assume that this distance is estimated accurately or, if biases in depth perception impact accuracy, that these biases also impact the perceived speed of the ball in such a way that these biases cancel out. Out of these two scenarios, we believe the latter is more likely, as it is quite established in the literature that depth is underestimated in VR [41], and an underestimation of depth would lead to an underestimation of the physical distance between the point of disappearance and the target rectangle. However, this bias in depth perception should also lead the observer to underestimate the physical speed of the ball in the same way, causing both biases to cancel each other out. Between-participant variability is neglected here.

Prediction

For the prediction task, under the above assumptions, participants are expected to respond between 0.12 and 0.2s earlier during the Opposite Direction motion profile than during the Static motion profile (Fig 3A). Our model further predicts that visual self-motion during motion observation will lead to higher variability in responses. Measuring the relation between self-motion and variability is not straight-forward because self-motion should cause an underestimation of the occlusion duration. If most noise behaves according to Weber’s Law, noise should be proportional to the mean length of the extrapolated interval. A shorter predicted interval should thus in turn be related to lower variability (in absolute terms) even if self-motion has no direct effect on precision. Fig 3B illustrates the expected relationship between biases in prediction, the motion profile, and variability in responses.

Fig 3.

Fig 3

A. Predicted data for the timing error in the prediction task, divided up by occlusion durations (x axis) and motion profile (color-coded; left-most: “Observer Static”; in the middle: “Opposite Directions”; right-most “Same Directions”). B. Predicted data for variability in the prediction task. The y axis displays the standard deviation of the extrapolated duration per condition and participant, while the x axis corresponds to the mean of the extrapolated duration per condition and participant. The motion profile is coded with different colors and line types (red and continuous for “Observer Static”, yellow and dashed for “Opposite Directions” and blue and dashed-and-dotted for “Same Direction”). The lines are regression lines for their respective condition.

Speed estimation

Here, we expect to replicate the findings from our previous study [30]: There we found that participants largely estimated speed with the same degree of accuracy when they were static as when they were moving in the same direction as the target. In line with these results, visually simulated self-motion in the opposite direction to the ball should lead to an overestimation of ball speed (Hypothesis 2a; see Fig 4A). Since we use a higher self-motion speed than in our previous study, we also expect that precision will be lower for visual self-motion in the opposite direction to the ball (Hypothesis 2b, see Fig 4B).

Fig 4.

Fig 4

A. Predicted PSEs (y axis) for each ball speed (x axis) and motion profile (color-coded; left-most: “Observer Static”; in the middle: “Opposite Directions”; right-most “Same Directions”). B. As A. but for the predicted JNDs.

A link between speed estimation and predicted time-to-contact

We further expect the errors observed in the prediction task in response to self-motion to correlate with the errors in the speed estimation task in response to self-motion, indicating that performance in speed perception translate to errors in predicted time-to-contact, both in terms of accuracy (Fig 5A) and precision (Fig 5B).

Fig 5.

Fig 5

A. Relationship between the difference in PSEs between the Opposite Directions motion profile and the Observer Static motion profile in the speed estimation task (x axis) and the difference in predicted durations between these motion profiles (y axis). One data point corresponds to one participant. B. As A., but for the relation between the JND differences in the speed estimation task between the “Opposite Directions” motion profile and the “Observer Static” motion profile and the differences in standard deviations between these motion profiles.

Data analysis plan

We first perform an outlier analysis. For the prediction task, we exclude all trials where the response timing was more than three times the occlusion duration, which indicates that the participant has not paid attention and missed the trial. For the speed estimation task, we exclude participants where more than 20% of presented ball cloud speeds were at the limits we set for the staircase (one third of the speed of the single ball and three times the speed of the single ball). For all analyses related to precision, we further exclude all conditions where we obtained a standard deviation of 0.01 or lower. According to our simulations, this should occur very rarely, and taking the log of such low values, as we do for the precision analyses to counteract the expected skew in these distributions, would lead to extremely small numbers that could bias results unduly. We also remove all trials where the head rotation was outside of the permitted range (+- 2.5°) for half or more of the recorded bins.

Unless noted otherwise, we compute bootstrapped 95% confidence intervals as implemented in the confint() function for base R [42] to determine statistical significance.

Prediction

To test Hypotheses 1a regarding accuracy, we use Linear Mixed Modelling as implemented in the lme4 package [43] for R. The corresponding script can be found here (on GitHub: https://github.com/b-jorges/Predicting-while-Moving/blob/main/Analysis%20Prediction.R). We fit a model with the temporal error as dependent variable, the motion profile (“Observer Static”, “Same Direction” and “Opposite Directions”) as fixed effect, and random intercepts and random slopes for the speed of the ball per participant, as well as random intercepts for the occlusion duration as random effects. In Wilkinson & Rogers notation (1973) [44], this model reads as follows:

ErrorMotionProfile+(SpeedBall|Participant)+(1|OcclusionDuration) (3)

We expect the regression coefficient corresponding to the motion profile “Opposite Directions” to be negative and significantly different from zero.

For Hypothesis 1b regarding precision, we need to take into account one possible confound: differences in timing accuracy can impact variability: overestimating the time it takes the ball to hit the rectangle could be connected to a higher variability, while underestimating the time could lead to lower variability. For this reason, we first compute the means and standard deviations of extrapolated durations for each condition and participant. We then fit a test model with the standard deviations as dependent variable, the mean timing error and the motion profile as fixed effects, and random intercepts as well as random slopes for ball speeds per participant and random intercepts for the occlusion durations as random effects:

log(SDofExtrapolatedTime)MeanofExtrapolatedTime+MotionProfile+(SpeedBall|Participant)+(1|OcclusionDuration) (4)

We further fit a null model without the motion profile as fixed effect:

log(SDofExtrapolatedTime)MeanofExtrapolatedTime+(SpeedBall|Participant)+(1|OcclusionDuration) (5)

We then compare both models by means of a Likelihood Ratio Test to determine whether the motion profile explains significantly more variability than the test model which already takes into account biases in extrapolated time. We will not interpret the regression coefficients as means and standard deviations are likely to be correlated, which may lead to biased regression coefficients.

Speed estimation

To test Hypotheses 2a and 2b (script can be found here, on GitHub: https://github.com/b-jorges/Predicting-while-Moving/blob/main/Analysis%20Speed%20Estimation.R) regarding speed estimation, we first use the R package quickpsy [45] to fit psychometric functions to the speed estimation data, separately for each participant, speed and motion profile. Quickpsy fits cumulative Gaussian functions to the data by direct likelihood maximization. The means of the cumulative Gaussians correspond to the Points of Subjective Equality (PSEs) and their standard deviations correspond to the 84.1% Just Noticeable Differences (JNDs).

To assess whether the motion profile biased the perceived speed significantly, we fit a Linear Mixed Model with the PSEs as dependent variable, the self-motion profile as fixed effect, and random intercepts and random slopes for the ball speed per participant as random effects:

PSEMotionProfile+(SpeedBall|Participant) (6)

We expect that the regression coefficient for the motion profile “Opposite Directions” will be positive and significantly different from zero, indicating that speed is overestimated when observer and target move in opposite directions as compared to when the observer is static.

Regarding precision, the same considerations apply as for the prediction task: in addition to a direct effect of the self-motion profile, biases elicited by the different self-motion profiles can impact precision. For this reason, we use a model comparison-based approach similar to the one used above. Separately for the “Same Direction” and “Opposite Directions” motion profiles, we first fit a test model that contains the log JNDs as dependent variable, the self-motion profile and the PSEs as fixed effects, and random intercepts as well as random slopes for ball speed per participant as random effects.

log(JND)MotionProfile+PSE+(SpeedBall|Participant) (7)

We also fit a null model without the motion profile as fixed effect:

log(JND)PSE+(SpeedBall|Participant) (8)

Finally, we compare both models with a Likelihood Ratio Test and we expect the test model (Eq 7) to be a significantly better fit than the null model (Eq 8). As for the prediction task, we will not interpret the regression coefficients obtained in this analysis.

A link between speed estimation and prediction

To test Hypotheses 3a and 3b (script can be found here, on GitHub: https://github.com/b-jorges/Predicting-while-Moving/blob/main/Analysis%20Correlation.R), we first prepare the prediction data by computing means and standard deviations of the extrapolated time for each participant. We then calculate the difference in performance (mean and standard deviations for the prediction task and PSEs and JNDs for the speed estimation task) between the “Opposite Directions” motion profile and the “Observer Static” motion profile for both tasks for each participant.

For accuracy, we then determine to what extent PSE differences between the Opposite Direction motion profile and the Observer Static motion profile obtained in the speed estimation task predict the mean extrapolated time in the prediction task. For this purpose, we fit a Linear Model with the difference in mean motion extrapolation errors between the motion profiles as the dependent variable and the difference in PSEs between the motion profiles as the independent variable:

MeanDifferenceinExtrapolatedTimetoContact(OppositeStatic)PSEDifference(OppositeStatic) (9)

We expect that the regression coefficient for the fixed effect “PSE Difference (Opposite–Static)” will be negative and significantly different from zero, indicating that a stronger effect of self-motion on PSEs is linked to a larger effect of self-motion on the estimated time-to-contact.

For precision, the same complication as for Hypothesis 1b applies: A correlation between the effect of visual self-motion on the precision of speed estimation and on the precision of the predicted times-to-contact could be due to biases introduced by visual self-motion. If visual self-motion in the opposite direction, for example, leads to too-early responses, the extrapolated intervals become shorter. A shorter interval, in turn, would lead to higher precision. Therefore, to test whether the difference in precision observed in the speed estimation task was significantly related to the variability in the prediction task even after accounting for biases, we need to determine whether the effect of visual self-motion on JNDs predicts any variability beyond the variability that is already explained by the bias in motion extrapolation. To test this hypothesis, we first fit a test model with the variability difference between the “Opposite Directions” motion profile and the “Observer Static” motion profile in the prediction task as the dependent variable and the mean difference between these motion profiles and the difference in JNDs in the speed estimation task as the independent variables (as a measure of bias introduced by visual self-motion):

VariabilityDifferenceinExtrapolatedTime(OppositeStatic)MeanDifferenceinExtrapolatedTime(OppositeStatic)+JNDDifference(OppositeStatic) (10)

We also fit a null model without the JND difference as independent variable:

VariabilityDifferenceinExtrapolatedTime(OppositeStatic)MeanDifferenceinExtrapolatedTime(OppositeStatic) (11)

Then, we use a Likelihood Ratio Test to determine whether the test model (with the JND difference as fixed effect) was significantly better than the null model. We expect the test model (Eq 10) to be a significantly better fit than the null model (Eq 11). As above, the regression coefficient will not be interpreted.

Model fitting

These analyses only serve to demonstrate that performance in both tasks is related, but they don’t provide insight into how strongly they are related. We will therefore use the models outlined in the section “Modelling the Predictions”, and described more in detail in S1 Appendix, to fit parameters that capture biases and precision differences in perceived speed due to self-motion (which we had set to 20% on average to model our predictions and conduct the power analyses) in both tasks separately.

We will use a two-step approach to fit these parameters: since we expect biases to affect variability in performance but do not expect variability to affect biases, we first set the variability parameter (capturing the impact of self-motion on the variability of perceived speed) to zero and fit the accuracy parameter (capturing the impact of self-motion on mean perceived speed). To do so, we minimize the root median squared error between the observed mean difference in timing errors between baseline and the self-motion condition and the difference in timing error in the simulated dataset across all speeds and occlusion durations. We perform this optimization by using the Brent method [46] as implemented in the optimize function in base R. In the second step, we set the accuracy parameter for each participant to the one fitted in the first step and fit the precision parameter. Here, we minimize the root median squared error between the difference between the standard deviation of the observed difference in timing errors in the baseline condition and the self-motion condition and the respective simulated values. We use the same approach to obtain these parameters for the speed estimation task as well but use the observed and simulated PSEs (for accuracy) and JNDs (for precision).

Once we have obtained one accuracy parameter and one precision parameter for each participant and task, we perform a simple linear regression between the parameters fitted for the prediction task and the speed estimation task (separately for accuracy, see Eq 12, and precision, see Eq 13) to determine to what extent performance in one task is indicative of performance in the other:

BiasPredictionBiasSpeedEstimation (12)
PrecisionDifferencePredictionPrecisionDifferenceSpeedEstmation (13)

Since these parameters are scaled the same way for both tasks (the effect of self-motion on accuracy/precision as a fraction of presented self-motion speed), we expect regression coefficients of around 1 in both analyses. A value of above 1 would mean that the effect of self-motion is stronger in the prediction task than in the speed estimation task, and vice-versa. To test this prediction we compute 95% confidence intervals which we expect to contain a value of 1.

An effect of visual self-motion in the same direction as the ball

While our earlier results [30] suggest that visual self-motion in the same direction as the observer should not have any effect on perceived speed, we perform all analyses outlined in this section equivalently for the “Same Direction” motion profile as well.

Power analysis

Since power for complex hierarchical designs cannot be computed analytically, we used Monte Carlo simulations to determine power for all statistical tests outlined in the previous section: we used our models for the prediction task and the speed estimation task to first simulate full datasets. Then, we performed the analyses detailed above over each of these simulated datasets and determined the results for each combination of number of participants and number of trials. To keep the computational cost manageable, we used the faster, but more bias-prone Satterthwaite approximation, as implemented in the lmerTest package [47] for R, to assess statistical significance rather than bootstrapped confidence intervals. The script used for the power analyses can be found here (on GitHub: https://github.com/b-jorges/Predicting-while-Moving/blob/main/Power%20Analysis.R).

We repeated this process 250 times for all combinations of 20, 30 and 40 participants, 5, 9 and 13 repetitions per condition in the prediction task, and 20 to 27, 30 to 37, and 40 to 47 trials per pest, which makes for an average of 50, 70 and 90 trials per condition, respectively, for the speed estimation task. The results are shown in Fig 6. For the precision in the prediction task, using 9 repetitions per condition appears to add a considerable amount more power than using only 5 repetitions, while the added benefit of another 4 repetitions for a total of 13 is small. However, the prediction task is very quick to do, taking only around 10 minutes even with 13 repetitions per condition. Similarly, 70 trials per condition increases the power to detect an effect on precision significantly more than using only 50 trials, while the added benefit of 90 trials is marginal. Since the speed estimation task takes much longer and is more fatiguing than the prediction task, we judge this marginal increase in power to be not worth the additional time spent by the participant. We thus opt for a combination of 40 participants, 13 repetitions per condition in the prediction task, and 70 trials per condition in the speed estimation task, which allows us to achieve a power of at least 0.85 for all statistical tests.

Fig 6.

Fig 6

Simulated power for the prediction task (A), the speed estimation task (B) and the correlation between performance in speed estimation and speed prediction (C), separately for the statistical tests referring to biases (accuracy) and variability (precision). The number of participants for which we simulated power is on the x axis, while the number of trials for each task is coded with different shades of green and line types. The horizontal lines indicate a power level of 0.8, 0.9 and 0.95 respectively.

We also used our power analyses to determine that all of our statistical analyses led to an expected false positive rate of 0.05 in absence of a true effect.

Supporting information

S1 Appendix

(DOCX)

Data Availability

All scripts and all data obtained with regards to this project will be made available in the project GitHub repository (https://github.com/b-jorges/Predicting-while-Moving/). Larger files such as the programs used to present stimuli and the respective Unity projects are hosted on OSF (https://osf.io/eayf7/).

Funding Statement

BJ and LRH are supported by the Canadian Space Agency (CSA) (CSA: 15ILSRA1-York). LRH is supported by a Discovery Grant from the Natural Sciences and Engineering Research Council (NSERC) of Canada (NSERC: RGPIN-2020-06093). The funders did not play any role in the study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Schmolesky MT, Wang Y, Hanes DP, Thompson KG, Leutgeb S, Schall JD, et al. Signal timing access the macaque visual system. J Neurophysiol. 1998;79: 3272–3278. doi: 10.1152/jn.1998.79.6.3272 [DOI] [PubMed] [Google Scholar]
  • 2.Maunsell JHR, Gibson JR. Visual response latencies in striate cortex of the macaque monkey. J Neurophysiol. 1992;68: 1332–1344. doi: 10.1152/jn.1992.68.4.1332 [DOI] [PubMed] [Google Scholar]
  • 3.DiCarlo JJ, Maunsell JHR. Using neuronal latency to determine sensory-motor processing pathways in reaction time tasks. J Neurophysiol. 2005;93: 2974–2986. doi: 10.1152/jn.00508.2004 [DOI] [PubMed] [Google Scholar]
  • 4.de Rugy A, Marinovic W, Wallis G. Neural prediction of complex accelerations for object interception. J Neurophysiol. 2012;107: 766–771. doi: 10.1152/jn.00854.2011 [DOI] [PubMed] [Google Scholar]
  • 5.Zago M, McIntyre J, Senot P, Lacquaniti F. Internal models and prediction of visual gravitational motion. Vision Res. 2008;48: 1532–1538. doi: 10.1016/j.visres.2008.04.005 [DOI] [PubMed] [Google Scholar]
  • 6.Foxe JJ, Simpson G V. Flow of activation from V1 to frontal cortex in humans: A framework for defining “early” visual processing. Exp Brain Res. 2002;142: 139–150. doi: 10.1007/s00221-001-0906-7 [DOI] [PubMed] [Google Scholar]
  • 7.Jörges B, López-Moliner J. Earth-Gravity Congruent Motion Facilitates Ocular Control for Pursuit of Parabolic Trajectories. Sci Rep. 2019;9: 1–13. doi: 10.1038/s41598-019-50512-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Spering M, Schutz AC, Braun DI, Gegenfurtner KR. Keep your eyes on the ball: smooth pursuit eye movements enhance prediction of visual motion. J Neurophysiol. 2011;105: 1756–1767. doi: 10.1152/jn.00344.2010 [DOI] [PubMed] [Google Scholar]
  • 9.Kreyenmeier P, Fooken J, Spering M. Context effects on smooth pursuit and manual interception of a disappearing target. J Neurophysiol. 2017;118: 404–415. doi: 10.1152/jn.00217.2017 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Dupin L, Wexler M. Motion perception by a moving observer in a threedimensional environment. J Vis. 2013;13: 1–14. doi: 10.1167/13.2.15 [DOI] [PubMed] [Google Scholar]
  • 11.Warren PA, Rushton SK. Perception of object trajectory: Parsing retinal motion into self and object movement components. J Vis. 2007;7: 2. doi: 10.1167/7.11.2 [DOI] [PubMed] [Google Scholar]
  • 12.Rushton SK, Warren PA. Moving observers, relative retinal motion and the detection of object movement. Curr Biol. 2005;15: R542–R53. doi: 10.1016/j.cub.2005.07.020 [DOI] [PubMed] [Google Scholar]
  • 13.Warren PA, Rushton SK. Evidence for flow-parsing in radial flow displays. Vision Res. 2008;48: 655–663. doi: 10.1016/j.visres.2007.10.023 [DOI] [PubMed] [Google Scholar]
  • 14.Warren PA, Rushton SK. Perception of scene-relative object movement: Optic flow parsing and the contribution of monocular depth cues. Vision Res. 2009;49: 1406–1419. doi: 10.1016/j.visres.2009.01.016 [DOI] [PubMed] [Google Scholar]
  • 15.Wertheim AH. Perceiving motion: relativity, illusions and the nature of perception. Neth J Psychol. 2008;64: 119–125. doi: 10.1007/bf03076414 [DOI] [Google Scholar]
  • 16.Wertheim AH. Motion perception during self-motion: The direct versus inferential controversy revisited. BBS 17:293–355. Behav Brain Sci. 1999;22: 337–340. doi: 10.1017/S0140525X99001831 [DOI] [Google Scholar]
  • 17.Layton OW, Niehorster DC. A model of how depth facilitates scenerelative object motion perception. PLoS Comput Biol. 2019;15: 1–29. doi: 10.1371/journal.pcbi.1007397 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Dupin L, Wexler M. Motion perception by a moving observer in a threedimensional environment. J Vis. 2013;13: 1–14. doi: 10.1167/13.2.15 [DOI] [PubMed] [Google Scholar]
  • 19.Warren PA, Rushton SK. Perception of object trajectory: Parsing retinal motion into self and object movement components. J Vis. 2007;7: 1–11. doi: 10.1167/7.11.2 [DOI] [PubMed] [Google Scholar]
  • 20.Dokka K, MacNeilage PR, DeAngelis GC, Angelaki DE. Multisensory self-motion compensation during object trajectory judgments. Cereb Cortex. 2015;25: 619–630. doi: 10.1093/cercor/bht247 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Fetsch CR, Deangelis GC, Angelaki DE. Visual-vestibular cue integration for heading perception: Applications of optimal cue integration theory. Eur J Neurosci. 2010;31: 1721–1729. doi: 10.1111/j.1460-9568.2010.07207.x [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Garzorz IT, Freeman TCA, Ernst MO, MacNeilage PR. Insufficient compensation for self-motion during perception of object speed: The vestibular Aubert-Fleischl phenomenon. J Vis. 2018;18: 1–9. doi: 10.1167/18.13.9 [DOI] [PubMed] [Google Scholar]
  • 23.Probst T, Loose R, Niedeggen M, Wist ER. Processing of visual motion direction in the fronto-parallel plane in the stationary or moving observer. Behav Brain Res. 1995;70: 133–144. doi: 10.1016/0166-4328(95)80003-4 [DOI] [PubMed] [Google Scholar]
  • 24.Gray R, MacUga K, Regan D. Long range interactions between object-motion and self-motion in the perception of movement in depth. Vision Res. 2004;44: 179–195. doi: 10.1016/j.visres.2003.09.001 [DOI] [PubMed] [Google Scholar]
  • 25.Dyde RT, Harris LR. The influence of retinal and extra-retinal motion cues on perceived object motion during self-motion. J Vis. 2008;8. doi: 10.1167/8.14.5 [DOI] [PubMed] [Google Scholar]
  • 26.Hogendoorn H, Alais D, MacDougall H, Verstraten FAJ. Velocity perception in a moving observer. Vision Res. 2017;138: 12–17. doi: 10.1016/j.visres.2017.06.001 [DOI] [PubMed] [Google Scholar]
  • 27.Niehorster DC, Li L. Accuracy and tuning of flow parsing for visual perception of object motion during self-motion. Iperception. 2017;8: 1–18. doi: 10.1177/2041669517708206 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Butler JS, Campos JL, Bülthoff HH. Optimal visual–vestibular integration under conditions of conflicting intersensory motion profiles. Exp Brain Res. 2014;233: 587–597. doi: 10.1007/s00221-014-4136-1 [DOI] [PubMed] [Google Scholar]
  • 29.Campos JL, Butler JS, Bülthoff HH. Multisensory integration in the estimation of walked distances. Exp Brain Res. 2012;218: 551–565. doi: 10.1007/s00221-012-3048-1 [DOI] [PubMed] [Google Scholar]
  • 30.Jörges B, Harris LR. Object speed perception during lateral visual self-motion. Attention, Perception, Psychophys. 2021;84: 25–46. doi: 10.3758/s13414-021-02372-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Jörges B, López-Moliner J. Determining Mean and Standard Deviation of the Strong Gravity Prior through Simulations. 2020; 1–19. doi: 10.31234/osf.io/exp93 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.de’Sperati C, Thornton IM. Motion prediction at low contrast. Vision Res. 2019;154: 85–96. doi: 10.1016/j.visres.2018.11.004 [DOI] [PubMed] [Google Scholar]
  • 33.Kammers MPM, van der Ham IJM, Dijkerman HC. Dissociating body representations in healthy individuals: Differential effects of a kinaesthetic illusion on perception and action. Neuropsychologia. 2006;44: 2430–2436. doi: 10.1016/j.neuropsychologia.2006.04.009 [DOI] [PubMed] [Google Scholar]
  • 34.Witt JK, Proffitt DR. Perceived slant: A dissociation between perception and action. Perception. 2007;36: 249–257. doi: 10.1068/p5449 [DOI] [PubMed] [Google Scholar]
  • 35.Bradshaw MF, Watt SJ. A dissociation of perception and action in normal human observers: The effect of temporal-delay. Neuropsychologia. 2002;40: 1766–1778. doi: 10.1016/s0028-3932(02)00039-8 [DOI] [PubMed] [Google Scholar]
  • 36.Zago M, McIntyre J, Senot P, Lacquaniti F. Visuo-motor coordination and internal models for object interception. Exp Brain Res. 2009;192: 571–604. doi: 10.1007/s00221-008-1691-3 [DOI] [PubMed] [Google Scholar]
  • 37.Dichgans J, Wist E, Diener HC, Brandt T. The Aubert-Fleischl phenomenon: A temporal frequency effect on perceived velocity in afferent motion perception. Exp Brain Res. 1975;23: 529–533. doi: 10.1007/BF00234920 [DOI] [PubMed] [Google Scholar]
  • 38.Taylor MM, Creelman CD. PEST: Efficient Estimates on Probability Functions. J Acoust Soc Am. 1967;41: 782–787. doi: 10.1121/1.1910407 [DOI] [Google Scholar]
  • 39.Kaiser MK. Angular velocity discrimination. Percept Psychophys. 1990;47: 149–156. doi: 10.3758/bf03205979 [DOI] [PubMed] [Google Scholar]
  • 40.Norman JF, Todd JT, Perotti VJ, Tittle JS. The Visual Perception of Three-Dimensional Length. J Exp Psychol Hum Percept Perform. 1996;22: 173–186. doi: 10.1037//0096-1523.22.1.173 [DOI] [PubMed] [Google Scholar]
  • 41.Jamiy F El, Marsh R. Distance estimation in virtual reality and augmented reality: A survey. IEEE Int Conf Electro Inf Technol. 2019;2019-May: 063–068. doi: 10.1109/EIT.2019.8834182 [DOI] [Google Scholar]
  • 42.R Core Team. A Language and Environment for Statistical Computing. R Foundation for Statistical Computing,. Vienna, Austria; 2017. Available: http://www.r-project.org/. [Google Scholar]
  • 43.Bates D, Mächler M, Bolker BM, Walker SC. Fitting linear mixed-effects models using lme4. J Stat Softw. 2015;67. doi: 10.18637/jss.v067.i01 [DOI] [Google Scholar]
  • 44.Wilkinson GN, Rogers CE. Symbolic Description of Factorial Models for Analysis of Variance. J R Stat Soc. 1973;22: 392–399. [Google Scholar]
  • 45.Linares D, López-Moliner J. quickpsy: An R Package to Fit Psychometric Functions for Multiple Groups. R J. 2016;8: 122–131. Available: https://journal.r-project.org/archive/2016-1/linares-na.pdf. [Google Scholar]
  • 46.Brent RP. Algorithms for Minimization Without Derivatives. IEEE Trans Automat Contr. 1974;19: 632–633. doi: 10.1109/TAC.1974.1100629 [DOI] [Google Scholar]
  • 47.Kuznetsova A, Brockhoff PB, Christensen RHB. lmerTest Package: Tests in Linear Mixed Effects Models. J Stat Softw. 2017;82. doi: 10.18637/jss.v082.i13 [DOI] [Google Scholar]

Decision Letter 0

Michael J Proulx

13 Feb 2022

PONE-D-21-37478The Impact of Visually Simulated Self-Motion on predicting object Motion – A Registered Report ProtocolPLOS ONE

Dear Dr. Jörges,

Thank you for submitting your Registered Report Protocol to PLOS ONE. After careful consideration, we feel that it has merit but requires some revision to meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. All of the reviewers share my enthusiasm for the potential of your study, and each made reasonable requests, suggests, and queries to help improve it so it will be suitable for acceptance upon completion. (Please note some of the auto-populated text provided by PLOS ONE might sound strange, as much of it is more suitable for a normal manuscript. It was clear to me and the reviewers what we were assessing here!)

Please submit your revised manuscript by Mar 30 2022 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Michael J Proulx, Ph.D.

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Thank you for stating the following in the Acknowledgments Section of your manuscript:

“BJ and LRH are supported by the Canadian Space Agency (CSA) (CSA: 15ILSRA1-York). LRH is supported by a Discovery Grant from the Natural Sciences and Engineering Research Council (NSERC) of Canada (NSERC: RGPIN-2020-06093). The funders did not play any role in the study design, data collection and analysis, decision to publish, or preparation of the manuscript.”

Please note that funding information should not appear in the Acknowledgments section or other areas of your manuscript. We will only publish funding information present in the Funding Statement section of the online submission form.

Please remove any funding-related text from the manuscript and let us know how you would like to update your Funding Statement. Currently, your Funding Statement reads as follows:

“BJ and LRH are supported by the Canadian Space Agency (CSA) (CSA: 15ILSRA1-York). LRH is supported by a Discovery Grant from the Natural Sciences and Engineering Research Council (NSERC) of Canada (NSERC: RGPIN-2020-06093). The funders did not play any role in the study design, data collection and analysis, decision to publish, or preparation of the manuscript.”

Please include your amended statements within your cover letter; we will change the online submission form on your behalf.

3. Please include your full ethics statement in the ‘Methods’ section of your manuscript file. In your statement, please include the full name of the IRB or ethics committee who approved or waived your study, as well as whether or not you obtained informed written or verbal consent. If consent was waived for your study, please include this information in your statement as well.

4. In your Data Availability statement, you have not specified where the minimal data set underlying the results described in your manuscript can be found. PLOS defines a study's minimal data set as the underlying data used to reach the conclusions drawn in the manuscript and any additional data required to replicate the reported study findings in their entirety. All PLOS journals require that the minimal data set be made fully available. For more information about our data policy, please see http://journals.plos.org/plosone/s/data-availability.

Upon re-submitting your revised manuscript, please upload your study’s minimal underlying data set as either Supporting Information files or to a stable, public repository and include the relevant URLs, DOIs, or accession numbers within your revised cover letter. For a list of acceptable repositories, please see http://journals.plos.org/plosone/s/data-availability#loc-recommended-repositories. Any potentially identifying patient information must be fully anonymized.

Important: If there are ethical or legal restrictions to sharing your data publicly, please explain these restrictions in detail. Please see our guidelines for more information on what we consider unacceptable restrictions to publicly sharing data: http://journals.plos.org/plosone/s/data-availability#loc-unacceptable-data-access-restrictions. Note that it is not acceptable for the authors to be the sole named individuals responsible for ensuring data access.

We will update your Data Availability statement to reflect the information you provide in your cover letter.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Does the manuscript provide a valid rationale for the proposed study, with clearly identified and justified research questions?

The research question outlined is expected to address a valid academic problem or topic and contribute to the base of knowledge in the field.

Reviewer #1: Partly

Reviewer #2: Yes

Reviewer #3: Yes

**********

2. Is the protocol technically sound and planned in a manner that will lead to a meaningful outcome and allow testing the stated hypotheses?

The manuscript should describe the methods in sufficient detail to prevent undisclosed flexibility in the experimental procedure or analysis pipeline, including sufficient outcome-neutral conditions (e.g. necessary controls, absence of floor or ceiling effects) to test the proposed hypotheses and a statistical power analysis where applicable. As there may be aspects of the methodology and analysis which can only be refined once the work is undertaken, authors should outline potential assumptions and explicitly describe what aspects of the proposed analyses, if any, are exploratory.

Reviewer #1: Yes

Reviewer #2: Partly

Reviewer #3: Yes

**********

3. Is the methodology feasible and described in sufficient detail to allow the work to be replicable?

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

4. Have the authors described where all data underlying the findings will be made available when the study is complete?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception, at the time of publication. The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above and, if applicable, provide comments about issues authors must address before this protocol can be accepted for publication. You may also include additional comments for the author, including concerns about research or publication ethics.

You may also provide optional suggestions and comments to authors that they might find helpful in planning their study.

(Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The proposed experiment is solid. It is however not set up clearly in the introduction: why does this need to be done, what gap in the literature does it fill? It is unclear to me what we learn from this effort.

For the rest, I have minor comments:

First few sentences of abstract velocity and speed are both used intermixed, to my understanding they do not have the same meaning. Stick to velocity until you get specific about speed.

P2, lines 13-18, the modelling results by Layton & Niehorster, https://doi.org/10.1371/journal.pcbi.1007397, is of relevance here. Further relevant to this page is https://doi.org/10.1177/2041669517708206.

P2, line 32: is there a “we” missing in this sentence about commitment? More generally, I am wondering why this complicated paragraph is needed at all, or at least it could be built up differently. It has been shown previously that for static observers of simulated self-motion, flow parsing is incomplete, i.e., not all of the self-motion is removed from the retinal motion of the object when judging object motion (see e.g. several warren & Rushton work, Niehorster & Li, 2017). You can then choose to speculate on some reasons, but label them as such from the beginning, the start of the story is that flow parsing is incomplete. That makes this easier to read and sets up the story more directly.

P2-3: I think reference should be made to the work of Wertheim, e.g.

Wertheim, A. H. (1994). Motion perception during self motion: The direct versus inferential controversy revisited. Behavioral and Brain Sciences, 17(2), 293–311. https://doi.org/10.1017/S0140525X00034646

Wertheim, A. H. (2008). Perceiving motion: Relativity, illusions and the nature of perception. Netherlands Journal of Psychology, 64(3), 119–125. https://doi.org/10.1007/BF03076414

He has theories and experimental work about how visual reference signals indicating the velocity of the eyeballs in space affect perception of motions in that space, both in terms of threshold and noise. That seems very directly related here, and if I remember and understand his theory correctly, it yields the additional prediction that the opposite motion condition will yield less overestimation than the same motion direction conditions yields underestimation, because a subtractive JND comes into play in both cases.

The arrows in figure 1 are unclear to me, why is noise increasing three times along the flow parsing pathway?

I’m struggling a little bit with the word predicted. At least, it is not motion that is being predicted, it is time to contact that is being predicted. The motion is known, albeit presumably misperceived, and supposedly held constant to yield the time to contact estimate. This held constant assumption is critical for your analysis. This is sloppy around the end of page 3, for instance, but occurs throughout. Perhaps the term extrapolation fits better here?

Relatedly, p4, line 4: remove the word motion from motion prediction, then the sentence makes sense to me.

P4 line 6, estimates, judgments may be a more appropriate word? Or percepts?

P4, line 25 and other links: please print the actual link in the article text. This way I is preserved better, e.g. in case the article is printed.

P4, line 35: “you”: rewrite

P5: line 6-7: rather critical sentence is incomplete, what happens after occlusion duration?

P5: why a gaussian velocity profile and not just a constant? This seems to me to complicate the situation. Assuming incomplete flow parsing, the object to be judged will also be seen to accelerate and decelerate (or vice versa) in the two movement conditions. What object speed is then used for the judgment, some kind of (weighted) average?

Do participants receive instructions about head motion? What happens to their view of the virtual world when they move their head, does it counterrotate so that the virtual environment is perceived as rigid? Would head movement make your data harder to analyze / add additional unwanted variability to your study (e.g. it could conceivably differ between conditions, although I am not directly aware of studies suggesting that the lateral simulated self-motion will induce head motion).

P5, line 23: is your task doable at very short durations (0.1 and 0.2 s especially)?

I see that the ball casts a shadow on the ground. Is that a deliberate choice? Speed of the shadow over the tiles on the ground (relative speed between the two) is a direct cue to ball speed that could be used more straightforwardly than motion of the ball itself.

Speed estimation: describe more clearly that its only the balls in the ball cloud that move, not the whole scene.

Is 37 trials sufficient for a JND estimate? Simulation work by Prins on his Bayesian staircase suggested you need more like 100 or so, if I recall correctly.

Are participants able to fixate the cross when it is so close to the rapidly moving ball cloud?

Do you experience induced motion in the fixation cross during simulated self-motion? May the opposite direction of this induced motion confound your results?

Do you use the same participants for the two tasks? Results such as Niehorster & Li 2017 suggest there may be wide variation between participants in how complete flow parsing is. That should give you ample variability between participants to do a strong test of correlation for you H3.

P 7, line 18-22. Why you assume that flow parsing is incomplete only in the opposite motion condition, but not in the same motion condition? That needs to be justified. Same for the precision prediction.

P 8, assumption atop the page: and you assume the distance is perceived correctly and is not affected by background motion. Can you justify this? If both perceived v and d may vary between conditions, you have a problem.

Reviewer #2: The study aims to examine an interesting topic: how self-motion influences prediction when interacting with moving objects. The abstract mentions a clear prediction: that perceived object speed is likely to be biased and more variable during self-motion, because separating object motion from self-motion might give rise to systematic errors and must presumably give rise to more variability. The authors propose to study this by presenting virtual moving targets during simulated self-motion (or absence thereof). They examine judgments of target speed by having participants compare the speed of the target with that of a moving cloud of dots within a static window. They examine prediction by having participants press a button when a ball reaches a target. The ball is occluded before it reaches the target. The prediction mentioned in the abstract makes perfect sense, but I feel that it is a very weak prediction: that there will be a correlation between performance in the two tasks. Specifying what one expects to be correlated might change this. I think it is a bit trivial that performance on the two tasks across different speeds of self-motion is correlated, but maybe the authors are referring to correlations across participants within each value of self-motion. Otherwise, maybe it makes more sense to check whether the values are similar, rather than only whether they are correlated (as in de la Malla et al., 2018, Errors in interception can be predicted from errors in perception. Cortex 98, 49-59). I actually see a theoretical complication in interpreting the data. Since the self-motion presumably shifts the goal (the target rectangle) as much as it does the ball, why would you expect any bias in judging self-motion to influence the timing of the tap? I think this needs to be explained.

Another issue that needs justification is the use of a fixation point. Apart from making the task quite unnatural, it also introduces many complications. First of all, how will fixation be ensured. It is very difficult to keep fixating while making judgments about moving targets, and small periods of pursuit at critical moments might influence one’s judgments. Secondly, the participants might make several of the judgments with respect to the fixation point. The fixation point does not move with the simulated self-motion, so its motion relative to the surrounding also needs to be interpreted. It also provides a reference in time for the button press: the time it took the ball to reach/cross fixation. At the very least the authors should explain why they have a fixation point, and how this might influence their results. I would consider not requiring fixation.

When the authors write “the process of flow parsing should add noise and lead to object speed judgements being more variable during self-motion” they are actually making some assumptions. Although these assumptions are probably reasonable, I think the authors should be explicit about the details. Assuming that people use some kind of flow parsing mechanism to separate object motion from self-motion, they presumably also have to do so when there is no self-motion. Thus, the assumption is that speed judgments become more noisy when self-motion is faster, just as they become more noisy (at least in absolute terms; it could be a fixed Weber fraction) when the object moves faster. Being very explicit about the assumptions will help the reader follow the reasoning. It might also be worthwhile more explicitly considering the consequences of the visual self-motion information being in conflict with information from other sources. Following Figure 1, perceived self-motion should be weak because 3 of the 4 ‘senses’ of the multisensory integration indicate that there is no self-motion. There is no evident reason for an asymmetry between motion with or against the ball. I am also not so sure about this interpretation of ‘flow parsing’. Flow parsing refers to the ability to separate object motion from self-motion from the visual information alone. That is indeed necessary for the proposed processing, but I don’t think that a multisensory value of self-motion is normally considered as an input to flow parsing, so maybe the terminology should be adjusted here. Actually, many of the claims and assumptions do not appear to be necessary for answering the question as to whether biases in speed estimation give the anticipated errors in prediction, so probably the introduction (and methods) can be simplified. Moreover, the last pair of hypotheses are what the authors really want to test (I think). They need to check that their manipulation (simulated self-motion) influences judged object speed (and its variability) but actually they already know that it will. Hypothesis 2 is therefore a bit superfluous. They plan to examine whether motion prediction is also influenced (Hypothesis 1) and whether it is influenced in the same manner (hypothesis 3). If it is influenced in the same manner, it must be influenced, so hypothesis 1 is also superfluous. This gives a much clearer study with one hypothesis (with two components: bias and variability).

There are also a number of things to consider in the methods. Especially if people will be tested at home, the authors might want to consider the extent to which participants are allowed to move their heads, and whether such head movements will be compensated for.

Why are stereoblind participants excluded? Do the authors expect their performance to be different? Is it a good idea to always center the trajectory in front of the observer, especially when that position is indicated by a fixation point? Maybe the authors should consider adding some jitter to the position. Otherwise the task could be performed by pressing the button after the same time from when the ball reaches fixation as the time between the ball appearing and it reaching fixation.

It appears to me from the video that the target disappears when the participant presses the button. Is that correct? This should be mentioned explicitly. Since the task is to press the button when the ball would hit the target, this task could be interpreted as judging the time of collision of two moving items, rather than in terms of self-motion. If the target’s motion is underestimated due to motion in the surrounding one might therefore find no effect even though the hypothesis is true. Is there some reason to exclude this possibility? Why was this velocity profile chosen for the self-motion? Not having a constant speed means that the response could be different for the two tasks simply because the moment that is considered relevant is different: for judging speed, presumably only the average speed is relevant, whereas for prediction the change in speed is presumably also relevant. I assume that the training on the prediction task was always with the observer static. This should be specified. Why is there no target in the speed estimation task (in the condition with a single ball)? Might this not influence the comparison? Nice instruction video! I assume you also have a version with the other order.

The status of the assumptions in the predictions section is not quite clear. Some of the assumptions are predictions based on earlier findings, but if the current results turn out to be slightly different it is not a problem. For instance, if the speed of the ball is overestimated by 30% rather than 20% at this speed (or the Weber fraction is not 10%) the reasoning will still hold in the same manner. In the case of the variability it might even be a problem if the results were identical to the previous ones (no influence of self-motion). The third assumption is very philosophical. How would you know whether they have the same bias other than by comparing performance in the two tasks, which is what the study was planned to examine so it cannot be an assumption. The same is true for Equation 3. The Weber fraction of 5% for distance judgments is presumably really an assumption that must be considered when converting speed judgment uncertainty into temporal uncertainty using Equation 3. Maybe explain exactly how this is done and therefore how sensitive the result is to deviations from this value.

In the motion prediction section I think it would be a good idea to clarify that certain predictions are based on earlier research, while others are based on reasoning. This might be important for the interpretation, because not finding the asymmetrical influence of background motion, for instance, need not affect the general conclusion, whereas not finding an increase in the standard deviation with the magnitude would make some of the proposed analysis meaningless.

In Figure 5A I am guessing the y-axis should be in s, not m/s. Why do the authors anticipate precisely this relationship? I think the authors can be a bit more specific about the actual values. Presumably these duration values are obtained by multiplying the difference in PSE by the occlusion time, or something like that. I would be specific, because that is what makes pre-registration a powerful tool. Figure 5B also confused me. If the authors expect such a mess, why bother?

I am not very familiar with the Wilkinson & Rogers notation, so I may be wrong, but it appears from Equation 4 that the authors assume linear, independent effects of observer motion, ball motion and occlusion duration. Why? Would you not for instance expect a larger effect of speed for a longer occlusion duration? Just under that equation the authors speak of biases in timing error. Do they mean systematic errors? This is not really a bias but a potential finding: that observer motion influences timing errors. What would not finding such an effect mean? Maybe the target position is shifted to the same extent as the ball, so their effects cancel? I see many potentially interesting issues to explore, but the idea of pre-registration is to precisely specify what you are testing. For this, I think the authors need to better specify which effect they expect and why. For equations 5 and 6 the measure is clear: all that matters is whether including the Motion Profile in the model provides a significant improvement. Finally, it seems that equation 9 is evaluating whether the judged speed influences the judged timing. Is this really what you want to know? Should you not be testing whether differences in judged speed can fully account for the differences in timing? By equations 10 and 11 you lost me completely. If the time difference and the JND difference are not independent this might give confusing results.

In the power analysis I do not see any measure of the original assumed variability and effect size. Maybe I missed something.

Reviewer #3: Overall, I believe that your study covers an interesting and important topic. It is well designed and the hypotheses are clear and well based on previous literature. I have some suggestions for improvement listed below.

On Page 2, line 21 you mention that “in many virtual reality (VR) applications, vestibular and proprioceptive cues signal that the body is at rest, while the visual optic flow cues simultaneously indicate self-motion.” Could you provide some examples and also explain why this is the case just in some VR applications but not others (e.g. is it due to properties of the hardware or the virtual environment itself?).

On Page 3 lines 11-19 there seem to be references missing for some of the statements you make.

On Page 3, Line 24 you say “it would seem logical that the prediction reflects this bias in motion estimation”. I would like to see a more detailed explanation for this assumption since it is critical for your study. I find that entire paragraph containing explanations that are a bit rushed and unclear.

I greatly appreciate the attention given to the participant sample size and counterbalance of gender and order of conditions.

I am however not sure of the acceptance of any VR HMDs owned by participants. You mention that in-person testing would be conducted on a VIVE Pro Eye if granted permission, but it is expected that participants may possess different HMDs such as Quest 1 or 2, which have significantly different specifications and most importantly, interaction methods (e.g. controllers). Especially for time-sensitive stimuli that you present, it would be important to first conduct a pre-assessment of how your Unity code runs on these HMDs. I understand that due to COVID restrictions currently in place you would have to test remotely, but I believe more should be done to mitigate potential limitations arising from this. Perhaps one option would be to cap the framerate and field of view to certain parameters which are compatible with those HMDs that are lowest in terms of specifications that you would still accept in your study.

For both tasks it is unclear how the speeds and sizes of the stimuli were determined. Was that based on previous literature? if so it should be mentioned or otherwise it should be based on piloting data.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

Reviewer #3: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2023 Jan 30;18(1):e0267983. doi: 10.1371/journal.pone.0267983.r002

Author response to Decision Letter 0


30 Mar 2022

Reviewer #1:

The proposed experiment is solid. It is however not set up clearly in the introduction: why does this need to be done, what gap in the literature does it fill? It is unclear to me what we learn from this effort.

(1.1) It is important to study prediction to know how humans deal with (a) occlusions and (b) neural delays in the transduction of motor commands. When studying how humans predict how objects will move in their environment, it is then important to know what information these predictions are based on. While it seems fairly likely that the same biases observed in speed estimation would also apply to the prediction of time-to-contact, we believe that this shouldn’t be assumed, and it has, to our knowledge, not been studied empirically. We have reworked the introduction to make the relevance of the proposed experiment clearer.

First few sentences of abstract velocity and speed are both used intermixed, to my understanding they do not have the same meaning. Stick to velocity until you get specific about speed.

(1.2) Thank you, addressed.

P2, lines 13-18, the modelling results by Layton & Niehorster, https://doi.org/10.1371/journal.pcbi.1007397, is of relevance here. Further relevant to this page is https://doi.org/10.1177/2041669517708206.

(1.3) Thank you, these are indeed highly pertinent additions.

P2, line 32: is there a “we” missing in this sentence about commitment? More generally, I am wondering why this complicated paragraph is needed at all, or at least it could be built up differently. It has been shown previously that for static observers of simulated self-motion, flow parsing is incomplete, i.e., not all of the self-motion is removed from the retinal motion of the object when judging object motion (see e.g. several warren & Rushton work, Niehorster & Li, 2017). You can then choose to speculate on some reasons, but label them as such from the beginning, the start of the story is that flow parsing is incomplete. That makes this easier to read and sets up the story more directly.

(1.4) We agree with the reviewer’s suggestions and have edited this paragraph accordingly.

P2-3: I think reference should be made to the work of Wertheim, e.g.

Wertheim, A. H. (1994). Motion perception during self motion: The direct versus inferential controversy revisited. Behavioral and Brain Sciences, 17(2), 293–311. https://doi.org/10.1017/S0140525X00034646

Wertheim, A. H. (2008). Perceiving motion: Relativity, illusions and the nature of perception. Netherlands Journal of Psychology, 64(3), 119–125. https://doi.org/10.1007/BF03076414

He has theories and experimental work about how visual reference signals indicating the velocity of the eyeballs in space affect perception of motions in that space, both in terms of threshold and noise. That seems very directly related here, and if I remember and understand his theory correctly, it yields the additional prediction that the opposite motion condition will yield less overestimation than the same motion direction conditions yields underestimation, because a subtractive JND comes into play in both cases.

(1.5) It appears to us that according to Wertheim’s account (as per his 2008 paper), object motion should be underestimated in both cases (for self-motion in the opposite direction and in the same direction as the object). In both cases, one JND worth of “reference motion” would be subtracted from the object motion signal, leading to an underestimation of object motion. This is not what we found in our previous paper (Jörges & Harris, 2021), where we saw evidence for an overestimation of speed for observer and target motion in opposite directions, and no effect for observer and target motion in opposite directions. Omitting the work of Wertheim was, nonetheless, an oversight. We have added these references and thank the reviewer for pointing it out to us.

The arrows in figure 1 are unclear to me, why is noise increasing three times along the flow parsing pathway?

(1.6) It is meant to convey that, qualitatively, the self-motion-based estimate is noisier than the optic flow estimate, which then carries over to the estimate of object velocity and the predicted time-to-contact. We removed the additional (and confusing) noise icons.

I’m struggling a little bit with the word predicted. At least, it is not motion that is being predicted, it is time to contact that is being predicted. The motion is known, albeit presumably misperceived, and supposedly held constant to yield the time to contact estimate. This held constant assumption is critical for your analysis. This is sloppy around the end of page 3, for instance, but occurs throughout. Perhaps the term extrapolation fits better here?

(1.7) “Extrapolation” is indeed more precise, and we have changed “motion prediction” to “motion extrapolation” throughout, or changed the wording to “prediction of time-to-contact” or similar.

Relatedly, p4, line 4: remove the word motion from motion prediction, then the sentence makes sense to me.

P4 line 6, estimates, judgments may be a more appropriate word? Or percepts?

P4, line 25 and other links: please print the actual link in the article text. This way I is preserved better, e.g. in case the article is printed.

P4, line 35: “you”: rewrite

(1.8) Thank you, we addressed these points.

P5: line 6-7: rather critical sentence is incomplete, what happens after occlusion duration?

(1.9) Addressed. This sentence is only meant to express that the distance between the point of disappearance and the target depends on the speed of the object and the occlusion duration.

P5: why a gaussian velocity profile and not just a constant? This seems to me to complicate the situation. Assuming incomplete flow parsing, the object to be judged will also be seen to accelerate and decelerate (or vice versa) in the two movement conditions. What object speed is then used for the judgment, some kind of (weighted) average?

(1.10) The intention behind a Gaussian motion profile was to generate a more natural stimulus that could also be used on a MOOG motion platform in later experiments with physical motion. However, we agree with the reviewer in that it introduces some unnecessary complexity and have changed the self-motion profile to a constant speed of 4 m/s.

Do participants receive instructions about head motion? What happens to their view of the virtual world when they move their head, does it counterrotate so that the virtual environment is perceived as rigid? Would head movement make your data harder to analyze / add additional unwanted variability to your study (e.g., it could conceivably differ between conditions, although I am not directly aware of studies suggesting that the lateral simulated self-motion will induce head motion).

(1.11) The program is designed such that head motion occurs indeed relative to the scene, e.g., theoretically the person could turn all the way around and the visual scene would be behind them. We assume that any head movements would add only random variability; however, a literature search has not revealed any studies conducted to support this assumption. We have therefore added a safeguard to the experiment: we will instruct them to keep their heads still and to ensure they do this we will display a circle in the middle of the field of view that moves with their head. The circle disappears when it is close enough to the fixation cross (+-2.5°), and trials will only start when this condition is met i.e., when their head is aligned. We will also record the participant’s head rotation throughout the experiment.

P5, line 23: is your task doable at very short durations (0.1 and 0.2 s especially)?

(1.12) While 0.1s and 0.2s are certainly with the range of reaction times for some participants, they can start planning their response before the ball disappears because the target rectangle is visible throughout the whole trial. This should allow them to complete the task successfully even at very short occlusion durations.

I see that the ball casts a shadow on the ground. Is that a deliberate choice? Speed of the shadow over the tiles on the ground (relative speed between the two) is a direct cue to ball speed that could be used more straightforwardly than motion of the ball itself.

(1.13) We opted to simulate the shadow to embed the ball more firmly in the visual scene, but it is true that the relative motion between shadow and tiles provide a more easily interpretable cue to object motion. We have therefore decided to remove the shadow.

Speed estimation: describe more clearly that its only the balls in the ball cloud that move, not the whole scene.

(1.14) We added a sentence to make this explicit.

Is 37 trials sufficient for a JND estimate? Simulation work by Prins on his Bayesian staircase suggested you need more like 100 or so, if I recall correctly.

(1.15) Please note that we run two staircases per condition, which brings the number of trials per condition closer to Prins’ suggestion. Further, adding more trials should only lower the variability of the measured JNDs per condition and participant. Since we are interested in the effect of our manipulation on the population level, we can counteract this increased variability on the condition-per-participant level by testing more participants. Our power analysis shows that a combination of 37 trials per PEST (i.e., 74 per condition) and 40 participants is sufficient.

Are participants able to fixate the cross when it is so close to the rapidly moving ball cloud?

(1.16) From our personal, subjective experience with the stimulus, the eye is sometimes drawn to the big moving ball, while the dot cloud is less problematic. However, we think that this is unlikely to differ between the self-motion conditions and tasks, and therefore it should not do more than add random (albeit unwanted) variability. Allowing the observer to view the scene freely is likely to lead them to pursue the target with their gaze, which may lead to an Aubert-Fleischl-induced underestimation of the target speed. This should also be similar across conditions and tasks, but would, in our opinion, just introduce more unwanted variability.

Do you experience induced motion in the fixation cross during simulated self-motion? May the opposite direction of this induced motion confound your results?

(1.17) In a previous study (Jörges & Harris, 2021) with a setup similar to the one in the proposed study, we found no evidence that a moving textured background induced motion in the target. While we did not use equivalence testing or a Bayesian analysis to confirm the absence of an effect, the sample size was relatively large (30 participants and around 50 trials per condition); that is, we expect any effect of induced motion should be very small. In contrast to our previous study, we also opted to keep the background blank (see Fig 2), precisely to minimize the risk of induced motion affecting performance.

Do you use the same participants for the two tasks? Results such as Niehorster & Li 2017 suggest there may be wide variation between participants in how complete flow parsing is. That should give you ample variability between participants to do a strong test of correlation for you H3.

(1.18) Yes, the same participants will complete both tasks, and thank you again for drawing our attention to the Nierhoster & Li paper.

P 7, line 18-22. Why you assume that flow parsing is incomplete only in the opposite motion condition, but not in the same motion condition? That needs to be justified. Same for the precision prediction.

(1.19) This is based on (surprising) previous findings (Jörges & Harris 2021): in this study, we only found differences in perceived speed when observer and ball were moving in opposite directions, while the perceived speed for observer motion in the same direction as the ball were similar to the static condition. Similarly, we saw descriptive, but not statistical differences in precision between the opposite directions and the static conditions, while precision in the same direction was again very similar to precision in the static condition.

P 8, assumption atop the page: and you assume the distance is perceived correctly and is not affected by background motion. Can you justify this? If both perceived v and d may vary between conditions, you have a problem.

(1.20) The presence of motion parallax as a cue during self-motion might lead participants indeed to estimate depth differently, which in turn might affect the perceived physical distance between the point of disappearance and the target in the prediction task. Depth is usually underestimated in VR, so the added motion parallax might make performance more accurate, i.e., participants might perceive the stimulus as further away during self-motion than when they are static. When the same retinal stimulation is interpreted as stemming from an object that is further away, its physical speed would be underestimated independently of whether object and observer move in the same or in opposite directions. This is not what we found in our previous study (Jörges & Harris, 2021): in the condition most similar to this experiment, we found an overestimation of speed when observer and object moved in opposite directions, and no discernible difference between a static observer and observer and object motion in the same direction, where we would have predicted an underestimation of object speed. While motion parallax might have some effect, its impact seems to be negligible when compared to other sources of variability.

Furthermore, since the task is presented continuously in the same environment, with the stimuli at the same, constant distance, it is in our opinion likely that participants build up one consistent depth estimate across all trials and all parts of the trajectory, which might be used as a prior to interpret optic flow information.

Finally, the participants come to a full halt in the moment the object disappears; that is, in principle they are able to judge the distance between the point of disappearance and the target in the prediction task while stationary in every self-motion condition.

Reviewer #2:

The prediction mentioned in the abstract makes perfect sense, but I feel that it is a very weak prediction: that there will be a correlation between performance in the two tasks. Specifying what one expects to be correlated might change this. I think it is a bit trivial that performance on the two tasks across different speeds of self-motion is correlated, but maybe the authors are referring to correlations across participants within each value of self-motion. Otherwise, maybe it makes more sense to check whether the values are similar, rather than only whether they are correlated (as in de la Malla et al., 2018, Errors in interception can be predicted from errors in perception. Cortex 98, 49-59).

(2.1) For Hypothesis 3, we are indeed looking for a correlation between performance in the speed estimation task and in the prediction task. However, we agree with the reviewer that merely finding a significant correlation between both tasks is a minimum requirement to support our hypothesis and perhaps not very informative on its own. To better assess not only if performance is correlated across both tasks, but also to which extent, we will use the models that generate our predictions to fit parameters that encode the effect of self-motion on perceived speed (separately for accuracy and precision) as a fraction of the self-motion speed. With that, they are scaled in the same way across both tasks, and any deviation from unity can be interpreted as either one task being affected more strongly by self-motion than the other. Reminiscent of de la Malla and her colleagues (2018), we can perform a regression between the fitted parameter in the prediction task and the fitted parameter in the speed estimation task and compute 95% confidence intervals, our prediction being that this confidence interval includes 1. To make this process more transparent to the reader, we have added a significant amount of detail information on the model and the fitting process in the manuscript.

I actually see a theoretical complication in interpreting the data. Since the self-motion presumably shifts the goal (the target rectangle) as much as it does the ball, why would you expect any bias in judging self-motion to influence the timing of the tap? I think this needs to be explained.

(2.2) Self-motion stops once the ball disappears, so in principle, participants should only be biased in estimating the speed of the object, not in estimating the distance between the point of disappearance and the target rectangle. One of the future avenues we see for this research is to introduce self-motion in different parts of the trial and compare to what extent the prediction of time-to-contact reflects biases in speed estimation (when self-motion occurs while the target is visible), in spatial updating (self-motion while the target is invisible), and both (self-motion throughout the whole trajectory).

Another issue that needs justification is the use of a fixation point. Apart from making the task quite unnatural, it also introduces many complications. First of all, how will fixation be ensured. It is very difficult to keep fixating while making judgments about moving targets, and small periods of pursuit at critical moments might influence one’s judgments. Secondly, the participants might make several of the judgments with respect to the fixation point. The fixation point does not move with the simulated self-motion, so its motion relative to the surrounding also needs to be interpreted. It also provides a reference in time for the button press: the time it took the ball to reach/cross fixation. At the very least the authors should explain why they have a fixation point, and how this might influence their results. I would consider not requiring fixation.

(2.3) When participants follow the target with their gaze, they might underestimate the target’s speed due to the Aubert-Fleischl phenomenon (c.f., Dichgans, Wist, Diener, & Brandt, 1975). While this shouldn’t affect the results in a way that confounds the experiment as it should only lower the perceived speed across all conditions, requiring fixation curtails this problem altogether. Since we want to have the option to collect data remotely, we can’t collect our participant’s eye-movements to ensure continued fixation and have to trust that our participants fixate as instructed. The fixation cross is always located right in front of the observer and does move with them during self-motion. We have made this clearer in the manuscript, and we have also provided the rationale for including the fixation cross.

We understand and agree with the concerns raised by the reviewer, but we believe that preventing Aubert-Fleischl-like effects from occurring outweighs these concerns.

When the authors write “the process of flow parsing should add noise and lead to object speed judgements being more variable during self-motion” they are actually making some assumptions. Although these assumptions are probably reasonable, I think the authors should be explicit about the details. Assuming that people use some kind of flow parsing mechanism to separate object motion from self-motion, they presumably also have to do so when there is no self-motion. Thus, the assumption is that speed judgments become more noisy when self-motion is faster, just as they become more noisy (at least in absolute terms; it could be a fixed Weber fraction) when the object moves faster. Being very explicit about the assumptions will help the reader follow the reasoning.

(2.4) Thank you and we made this more explicit in the introduction.

It might also be worthwhile more explicitly considering the consequences of the visual self-motion information being in conflict with information from other sources. Following Figure 1, perceived self-motion should be weak because 3 of the 4 ‘senses’ of the multisensory integration indicate that there is no self-motion.

(2.5) In our view, under most theories of multisensory integration this would depend on the weighting each cue receives (be it according to their relative reliabilities or other). We made a note of this in the introduction, but please note that we cut down on the section on multisensory perception of self-motion at the suggestion of reviewer #1, as we don’t have any commitment to a specific theory of multisensory integration of self-motion cues beyond self-motion speed being underestimated at least slightly when only visual cues are presented.

There is no evident reason for an asymmetry between motion with or against the ball.

(2.6) There is indeed no evident reason, but we have found evidence for such asymmetry that we consider rather compelling in a previous study (Jörges & Harris, 2021). While we will certainly verify statistically whether this asymmetry also holds for the prediction task (and is consistent between speed estimation and time-to-contact judgements), it seems reasonable to assume that the asymmetry will hold up for the purposes of planning the present study.

I am also not so sure about this interpretation of ‘flow parsing’. Flow parsing refers to the ability to separate object motion from self-motion from the visual information alone. That is indeed necessary for the proposed processing, but I don’t think that a multisensory value of self-motion is normally considered as an input to flow parsing, so maybe the terminology should be adjusted here.

(2.7) We have rephrased this section in the introduction and made clear that the original flow parsing hypothesis refers to a purely visual mechanism.

Actually, many of the claims and assumptions do not appear to be necessary for answering the question as to whether biases in speed estimation give the anticipated errors in prediction, so probably the introduction (and methods) can be simplified.

(2.8) At the reviewer’s (and reviewer #1’s) suggestion, we have removed some of the detail on the multisensory perception of self-motion and integrated it with the paragraph on flow parsing.

Moreover, the last pair of hypotheses are what the authors really want to test (I think). They need to check that their manipulation (simulated self-motion) influences judged object speed (and its variability) but actually they already know that it will. Hypothesis 2 is therefore a bit superfluous. They plan to examine whether motion prediction is also influenced (Hypothesis 1) and whether it is influenced in the same manner (hypothesis 3). If it is influenced in the same manner, it must be influenced, so hypothesis 1 is also superfluous. This gives a much clearer study with one hypothesis (with two components: bias and variability).

(2.9) We treat Hypothesis 2 partially as a replication, and partially as a more highly powered attempt to detect an effect of self-motion on variability. Hypothesis 1 aims to establish whether self-motion affects time-to-contact estimation, which seems likely, but shouldn’t be assumed a priori. Hypothesis 3 builds on Hypotheses 1 and 2, but could be true even if Hypotheses 1 and/or 2 are not supported: e.g., self-motion could affect participants in similar ways across both tasks, without there being a consistent population-wide effect in either direction. In this case, we would see the high correlation we expect if Hypothesis 3 is true without Hypotheses 1 and 2 being supported. It is also possible that participants show the expected effect of self-motion in both tasks on average, but that this effect is not consistent for each participant across both tasks. It is therefore our view that each of our hypotheses, while they of course build on each other and interact with each other, has merit of its own and should be treated separately.

There are also a number of things to consider in the methods. Especially if people will be tested at home, the authors might want to consider the extent to which participants are allowed to move their heads, and whether such head movements will be compensated for.

(2.10) Please see our response 1.11 to reviewer #1.

Why are stereoblind participants excluded? Do the authors expect their performance to be different?

(2.11) Stereovision is likely involved in judging the depth of the stimulus. While it seems most likely that stereo-blindness would only lead to unwanted, yet random variability rather than biases, administering stereotests can help reduce this variability at a very minor cost.

Is it a good idea to always center the trajectory in front of the observer, especially when that position is indicated by a fixation point? Maybe the authors should consider adding some jitter to the position. Otherwise the task could be performed by pressing the button after the same time from when the ball reaches fixation as the time between the ball appearing and it reaching fixation.

(2.12) Only the visible part of the trajectory is centered in front of the observer, and the target disappears by itself, without the observer’s intervention. We further use three different occlusion intervals (0.5, 0.6 and 0.7s), which should prevent the observer from circumventing the actual task by using heuristics.

It appears to me from the video that the target disappears when the participant presses the button. Is that correct? This should be mentioned explicitly. Since the task is to press the button when the ball would hit the target, this task could be interpreted as judging the time of collision of two moving items, rather than in terms of self-motion. If the target’s motion is underestimated due to motion in the surrounding one might therefore find no effect even though the hypothesis is true. Is there some reason to exclude this possibility?

(2.13) The target disappears after 0.5s of motion independently of the observer’s actions. The observer’s only task in the prediction experiment is to press the button when the ball would have hit the target rectangle (had it not disappeared). Self-motion occurs only while the ball is visible, that is, self-motion should not bias the estimated distance the ball has to cover between the point of disappearance and the target rectangle.

Why was this velocity profile chosen for the self-motion? Not having a constant speed means that the response could be different for the two tasks simply because the moment that is considered relevant is different: for judging speed, presumably only the average speed is relevant, whereas for prediction the change in speed is presumably also relevant.

(2.14) Please see our response 1.10 to reviewer #1.

I assume that the training on the prediction task was always with the observer static. This should be specified.

(2.15) The reviewer is correct in their assumption, and we have made this clearer in the manuscript.

Why is there no target in the speed estimation task (in the condition with a single ball)? Might this not influence the comparison?

(2.16) We added the target to the speed estimation task. You can watch a video (also included in the instruction video) here: https://youtu.be/AsROlBXzgr0

Nice instruction video! I assume you also have a version with the other order.

(2.17) Thank you, and we do, yes. We have added links to both versions to the paper.

The status of the assumptions in the predictions section is not quite clear. Some of the assumptions are predictions based on earlier findings, but if the current results turn out to be slightly different it is not a problem. For instance, if the speed of the ball is overestimated by 30% rather than 20% at this speed (or the Weber fraction is not 10%) the reasoning will still hold in the same manner. In the case of the variability it might even be a problem if the results were identical to the previous ones (no influence of self-motion). The third assumption is very philosophical. How would you know whether they have the same bias other than by comparing performance in the two tasks, which is what the study was planned to examine so it cannot be an assumption. The same is true for Equation 3. The Weber fraction of 5% for distance judgments is presumably really an assumption that must be considered when converting speed judgment uncertainty into temporal uncertainty using Equation 3. Maybe explain exactly how this is done and therefore how sensitive the result is to deviations from this value.

(2.18) These are assumptions that we make in order to simulate datasets, both to visualize what we expect the data to be collected to look like, and to conduct the power analyses. As such, we have to make assumptions about the effect size we are expecting to observe, like the assumption that the same biases and variability differences are at play in both tasks, that the perceived speed is biased by 20% of the self-motion speed, etc. Please note that our experiment is specifically intended to test these assumptions about the effect of self-motion on perceived speed. We have added a more detailed description of the way we model the expected data in Appendix A.

To address the reviewer’s sample question directly: Since there are no variability-weighted components in this model, deviations from the expected Weber Fractions should only lead to non-systematic changes in precision across all conditions.

In the motion prediction section I think it would be a good idea to clarify that certain predictions are based on earlier research, while others are based on reasoning. This might be important for the interpretation, because not finding the asymmetrical influence of background motion, for instance, need not affect the general conclusion, whereas not finding an increase in the standard deviation with the magnitude would make some of the proposed analysis meaningless.

(2.19) We have made some of the assumptions more explicit and marked clearly which are based on the literature, which relate to our hypotheses, and for which we have found support ourselves in our previous study (Jörges & Harris, 2021).

In Figure 5A I am guessing the y-axis should be in s, not m/s. Why do the authors anticipate precisely this relationship? I think the authors can be a bit more specific about the actual values. Presumably these duration values are obtained by multiplying the difference in PSE by the occlusion time, or something like that. I would be specific, because that is what makes pre-registration a powerful tool.

(2.20) Figure 5A shows the expected relationship between biases induced by self-motion in the speed estimation task (x axis) and the prediction task (y axis, which should indeed be in s rather than m/s – we thank the reviewer for spotting this), respectively. These predictions emanate from the models outlined under “Predictions”. For statistical testing, what matters is that we expect a negative relationship between the self-motion bias in the speed estimation task (operationalized as mean difference between the PSEs in the “Opposite Directions” condition and the “Static” condition) and the self-motion bias in the prediction task (operationalized as the mean difference between the timing error in the “Opposite Directions” self-motion condition and the ”Static” self-motion condition”).

As stated above, we agree that a mere (negative) correlation is a minimum finding, and we should indeed aim to quantify to what extent performance in one task predicts performance in the other. We have added a modelling section to the manuscript to address this issue.

Figure 5B also confused me. If the authors expect such a mess, why bother?

(2.21) The effect on precision (Figure 5B) looks indeed quite weak (visually). However, please note that the difference in variability underlying this plot is relatively big (an increase in variability of 20% of the presented speed). The issue here is that there are other sources of variability that, to some extent, obscure this effect in the measurement we take. We compensate for this high degree of noise by testing a much larger number of participants than we would test if we were only interested in the effect of self-motion on accuracy (see also the discrepancy between hypotheses relating to accuracy and hypotheses relating to precision in the power analysis).

I am not very familiar with the Wilkinson & Rogers notation, so I may be wrong, but it appears from Equation 4 that the authors assume linear, independent effects of observer motion, ball motion and occlusion duration. Why? Would you not for instance expect a larger effect of speed for a longer occlusion duration?

(2.22) By setting up the linear mixed models this way, we assume that each participant will show a linear effect of the ball speed on the timing error. We further assume no interactions between any of the effects. We agree that the assumption that observer motion, ball speed and occlusion direction interact is reasonable. However, the idea behind the random effects in a multilevel regression model (i.e., the effects specified in brackets) is to capture as much variability as possible in order to prevent uncaptured variability from biasing the regression coefficients of interest (in our case the regression coefficients pertaining to observer motion) and to raise statistical power. While it is generally desirable to match the expected structure of the variability as closely as possible, a large number of random effects (like a three-way interaction between observer motion, ball speed, and occlusion duration) can lead the model to become hard or even impossible to fit to the data. The structure we have chosen is a compromise between capturing the most important variability components and being able to fit the model. We have also verified its adequacy by performing a “power” analysis (or more accurately a “false positive analysis”) under the assumption of no effect, which yielded a false positive rate at the expected level of alpha = 0.05.

Just under that equation the authors speak of biases in timing error. Do they mean systematic errors? This is not really a bias but a potential finding: that observer motion influences timing errors.

(2.23) Indeed, this is one of the statistical hypotheses we are testing. It seems to us that our use of the term “bias” is equivalent to what the reviewer understands as “systematic errors”: mean differences in response to a manipulation, or differences in accuracy. This is the conventional use of the term “bias” in studies of this kind.

What would not finding such an effect mean? Maybe the target position is shifted to the same extent as the ball, so their effects cancel?

(2.24) Not finding an effect of the self-motion profile in this analysis would, per se, provide evidence that participants can extrapolate motion with a reasonable degree of accuracy even when they experience visual self-motion while they observe the object. What this means then depends on our findings for the other hypotheses: if we replicate our earlier findings (Jörges & Harris, 2021) and find that the participants in this study also overestimate the speed of objects they observe while experiencing self-motion in the opposite direction, then a negative result for hypothesis (1a) would mean that participants use other sources of information in the prediction task than in the speed estimation task (e.g., prior information from “Static” trials) or perhaps that information is processed differently in a purely perceptual task than in an action-oriented task. This would be a surprising result that we would discuss thoroughly in the discussion, and which would certainly call for further investigation.

In our opinion, the specific alternate explanation the reviewer mentions, namely that the effects of self-motion on perceived ball speed and target position might cancel out, is not quite as likely as the observer comes to a halt in the moment in which the ball disappears, such that they should be able to estimate the distance between the point of disappearance and the target rectangle accurately.

I see many potentially interesting issues to explore, but the idea of pre-registration is to precisely specify what you are testing. For this, I think the authors need to better specify which effect they expect and why.

(2.25) We expect the regression coefficient for the self-motion profile “Opposite Directions” to be significantly different from zero and positive, which would constitute evidence that the timing error is more negative (i.e., participants pressed the button too early) in the “Opposite Directions” self-motion profile than when participants are static. We have made our expectations explicit for all statistical tests mentioned in the analysis section of the manuscript.

For equations 5 and 6 the measure is clear: all that matters is whether including the Motion Profile in the model provides a significant improvement. Finally, it seems that equation 9 is evaluating whether the judged speed influences the judged timing. Is this really what you want to know? Should you not be testing whether differences in judged speed can fully account for the differences in timing?

(2.26) We agree with that reviewer in that the statistical test expressed in equation 9 – which assesses whether differences in judged speed relate to timing errors at all – is a very low bar, and it would be more interesting to assess the degree to which perceived speed explains timing differences. We do this by modelling the results (see new sub-section on “Model fitting”).

By equations 10 and 11 you lost me completely. If the time difference and the JND difference are not independent this might give confusing results.

(2.27) We are indeed almost certain that mean differences in timing and JND differences will share a relevant amount of variability. However, our analysis establishes whether JND differences explain any variability in the precision of the timing responses beyond what is already explained by the mean differences in timing. We will only use the Likelihood Ratio Test as evidence here and we will not interpret the regression coefficients obtained in this analysis as they are, as the reviewer points out, are impossible to interpret. We have made this explicit in the analysis section.

In the power analysis I do not see any measure of the original assumed variability and effect size. Maybe I missed something.

(2.28) For complex nested designs such as ours (with several participants completing several repetitions of several conditions), analytic power analyses (such as the ones implemented in the program G Power, for example) are not possible. For this reason, we used the models outlined under “Predictions” to generate synthetic datasets based on what we have called “assumptions” in that section (e.g., the different within- and between-participant sources of variability that are likely to influence the dependent variable in the task, as well as the strength of the effect of self-motion on both accuracy and precision). We simulated 250 datasets for each combination of number of participants and number of repetitions per condition and performed all analyses (see Equations 5 to 11) over these synthetic datasets. The fraction of datasets for which these analyses return a significant result can then be taken as the power for each analysis.

Reviewer #3:

On Page 2, line 21 you mention that “in many virtual reality (VR) applications, vestibular and proprioceptive cues signal that the body is at rest, while the visual optic flow cues simultaneously indicate self-motion.” Could you provide some examples and also explain why this is the case just in some VR applications but not others (e.g. is it due to properties of the hardware or the virtual environment itself?).

(3.1) At the suggestion of Reviewer #1, we reduced this paragraph and it no longer contains this reference to VR applications. We were referring to room-scale VR applications that allow the observer to walk around in the virtual environment on their own accord. But in fact any instance when you are watching a scene taken by a moving camera, e.g., on television or at the movies, provides this intersensory conflict.

On Page 3 lines 11-19 there seem to be references missing for some of the statements you make.

(3.2) While the first half of this paragraph is intended to draw the reader’s attention to some of the geometric properties of our setup, we added a reference (recommended by reviewer #1) that has explored some of these physical relationships (particularly those involving depth) and their impact on flow parsing.

On Page 3, Line 24 you say “it would seem logical that the prediction reflects this bias in motion estimation”. I would like to see a more detailed explanation for this assumption since it is critical for your study. I find that entire paragraph containing explanations that are a bit rushed and unclear.

(3.3) We have expanded on the processes at play (and possible scenarios in which predictions might not reflect these biases) in the introduction.

I am however not sure of the acceptance of any VR HMDs owned by participants. You mention that in-person testing would be conducted on a VIVE Pro Eye if granted permission, but it is expected that participants may possess different HMDs such as Quest 1 or 2, which have significantly different specifications and most importantly, interaction methods (e.g. controllers). Especially for time-sensitive stimuli that you present, it would be important to first conduct a pre-assessment of how your Unity code runs on these HMDs. I understand that due to COVID restrictions currently in place you would have to test remotely, but I believe more should be done to mitigate potential limitations arising from this. Perhaps one option would be to cap the framerate and field of view to certain parameters which are compatible with those HMDs that are lowest in terms of specifications that you would still accept in your study.

(3.4) We share the reviewer’s concerns and have limited the field of view in all programs to 80° (horizontally and vertically). Unity already caps the frame rate at 60 Hz, and drops below this framerate should occur almost never, as our programs require much less processing power than any VR applications the participants might use regularly. Our input method uses the keyboard for all participants to eliminate variability or technical issues due to different controllers.

For both tasks it is unclear how the speeds and sizes of the stimuli were determined. Was that based on previous literature? if so it should be mentioned or otherwise it should be based on piloting data.

(3.5) We have used a similar setup in our previous study (Jörges & Harris 2021), the only difference being that the single target and the ball in the ball cloud are now the same size. We made this change because we found that the ball cloud was judged consistently faster in the 2021 study, which we suspect to be a product of the difference in sizes between the single ball and the balls in the ball cloud in the previous study. The speeds and occlusion intervals were chosen such that, when participants keep their gaze on the fixation cross, the whole trajectory unfolds within a field of view of about 60° (of a total field of view of 80°), which is well within the field of view of any modern HMD. We have added these details in the methods section.

References:

Dichgans, J., Wist, E., Diener, H. C., & Brandt, T. (1975). The Aubert-Fleischl phenomenon: A temporal frequency effect on perceived velocity in afferent motion perception. Experimental Brain Research, 23(5), 529–533. https://doi.org/10.1007/BF00234920

Jörges, B., & Harris, L. R. (2021). Object speed perception during lateral visual self-motion. Attention, Perception, & Psychophysics.

Attachment

Submitted filename: Response to Reviewers #1.docx

Decision Letter 1

Michael J Proulx

20 Apr 2022

The Impact of Visually Simulated Self-Motion on Predicting Object Motion – A Registered Report Protocol

PONE-D-21-37478R1

Dear Dr. Jörges,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements (note some of this text might not be well adapted for Registered Reports, so bear that in mind). All three reviewers replied with Accept, and I am also satisfied with the changes. Well done. Note two reviewers have included a few other suggestions you might find useful.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Michael J Proulx, Ph.D.

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Does the manuscript provide a valid rationale for the proposed study, with clearly identified and justified research questions?

The research question outlined is expected to address a valid academic problem or topic and contribute to the base of knowledge in the field.

Reviewer #1: Yes

Reviewer #2: Yes

**********

2. Is the protocol technically sound and planned in a manner that will lead to a meaningful outcome and allow testing the stated hypotheses?

The manuscript should describe the methods in sufficient detail to prevent undisclosed flexibility in the experimental procedure or analysis pipeline, including sufficient outcome-neutral conditions (e.g. necessary controls, absence of floor or ceiling effects) to test the proposed hypotheses and a statistical power analysis where applicable. As there may be aspects of the methodology and analysis which can only be refined once the work is undertaken, authors should outline potential assumptions and explicitly describe what aspects of the proposed analyses, if any, are exploratory.

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Is the methodology feasible and described in sufficient detail to allow the work to be replicable?

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Have the authors described where all data underlying the findings will be made available when the study is complete?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception, at the time of publication. The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above and, if applicable, provide comments about issues authors must address before this protocol can be accepted for publication. You may also include additional comments for the author, including concerns about research or publication ethics.

You may also provide optional suggestions and comments to authors that they might find helpful in planning their study.

(Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The authors have done a very thorough and good job responding to my comments. I am looking forward to seeing the results of the study.

One further reference that may be of interest is https://jov.arvojournals.org/article.aspx?articleid=2770910

Reviewer #2: The purpose and methods are now much clearer. Most of my concerns clearly arose from having misunderstood details of the experiment. Many of the issues have been clarified sufficiently, but there are still a few small things that I believe will be helpful to clarify or motivate at some time.

The time course of the trials is now much clearer to me, which indeed solves many of my issues. I think it would be useful to modify figure 2C (and possibly D) to better match the actual experiment. This figure was the origin of much of my confusion, because in it the whole trajectory is more or less centred on fixation, rather than only the visible part. Probably the authors should also add something to illustrate when the observer moved (as now illustrated for the ‘invisible’ time). The pattern of events is quite complex, so it is good to have a reference. That might also help understand Table 1 which I still fail to understand, even when trying to work back from what it might mean. Is this the ball’s speed relative to fixation? I do not understand why the difference between ball speeds for the static observer changes in this manner. Or is there a typo somewhere? Should the first value be 20.4? Since the fixation point moves with the participant (but does not actually move) the ball moves in the opposite direction due to self-motion. That explains why opposite directions increases ball speed, but why this strange pattern for same speed? I feel that I am still missing something. It might help to always specify relative to what the motion is measured or described, because it is not intuitive. For instance, simulated self-motion does not correspond with retinal motion, because the participant is fixating a point that is moving along with the participant and is therefore static on the screen. The ball speed ‘should’ (according to the reasoning in the paper) be judged relative to the world, rather than relative to the observer. Are participants aware of this (is it part of the instruction)? In the timing task it is obvious because the self-motion affects the target, but in the speed judgment task this is not self-evident. All this could be problematic for the further interpretation, but not for the parts that are based on the authors’ previous work. I think being even clearer about the task and stimuli will make it easier for the reader to follow the reasoning.

Another issue that I had not always interpreted correctly is the role of the simulations and which parts of the methods are about simulations. I think it does make sense, but sometimes it is not clear to me whether the data presented in the figures are the outcome of the simulations, and sometimes it is not clear whether part of the analysis also applies to the simulations. I would try to clarify this. For instance, in the figure captions simply replace “Predicted data …” (Figure 3) by something like “Data from simulations based on …”. In the ‘data analysis plan’ indicate which parts (if any) also apply to the simulations. Maybe also change the order of some sections, because I think that the power analysis is based on the same (kind of) simulations as the predictions. Thinking logically I can guess what the authors did, but it is better to be told explicitly.

Details:

The order of the bars is incorrect in Figure 4: opposite directions in yellow (legend) but rightmost is same directions (caption). The power seems to be 0.75 for precision in speed estimation in Figure 6 (so less than 0.85).

I would explicitly mention that the fixation cross is static on the screen when mentioning that it moves with the observer (page 5 line 36). It is obvious when you think of it, but at this stage in the paper the reader does not need to think of this so it is worth pointing it out. In the next sentence I would also add the word ‘lateral’: The target is presented at a lateral distance that depends on the speed of the ball … The study is quite complex so it helps to guide the reader a bit.

The phrase “Given the on-going COVID-19 pandemic …” is probably no longer relevant.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

Acceptance letter

Michael J Proulx

25 Apr 2022

PONE-D-21-37478R1

The Impact of Visually Simulated Self-Motion on Predicting Object Motion – A Registered Report Protocol

Dear Dr. Jörges:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Michael J Proulx

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Appendix

    (DOCX)

    Attachment

    Submitted filename: Response to Reviewers #1.docx

    Data Availability Statement

    All scripts and all data obtained with regards to this project will be made available in the project GitHub repository (https://github.com/b-jorges/Predicting-while-Moving/). Larger files such as the programs used to present stimuli and the respective Unity projects are hosted on OSF (https://osf.io/eayf7/).


    Articles from PLOS ONE are provided here courtesy of PLOS

    RESOURCES