Abstract
Objective
We aim to bridge the gap between naturalistic studies of driver behavior and modern cognitive and neuroscientific accounts of decision making by modeling the cognitive processes underlying left-turn gap acceptance by human drivers.
Background
Understanding decisions of human drivers is essential for the development of safe and efficient transportation systems. Current models of decision making in drivers provide little insight into the underlying cognitive processes. On the other hand, laboratory studies of abstract, highly controlled tasks point towards noisy evidence accumulation as a key mechanism governing decision making. However, it is unclear whether the cognitive processes implicated in these tasks are as paramount to decisions that are ingrained in more complex behaviors, such as driving.
Results
The drivers’ probability of accepting the available gap increased with the size of the gap; importantly, response time increased with time gap but not distance gap. The generalized drift-diffusion model explained the observed decision outcomes and response time distributions, as well as substantial individual differences in those. Through cross-validation, we demonstrate that the model not only explains the data, but also generalizes to out-of-sample conditions.
Conclusion
Our results suggest that dynamic evidence accumulation is an essential mechanism underlying left-turn gap acceptance decisions in human drivers, and exemplify how simple cognitive process models can help to understand human behavior in complex real-world tasks.
Application
Potential applications of our results include real-time prediction of human behavior by automated vehicles and simulating realistic human-like behaviors in virtual environments for automated vehicles.
Keywords: driver behavior, decision making, computational modeling
INTRODUCTION
Understanding the seemingly mundane decisions made daily by millions of human drivers is critical for designing traffic infrastructure and engineering automated vehicles which can safely interact with humans around them. Overtaking a slow-moving vehicle, merging onto a highway, taking a left turn across path of an oncoming car—all these situations require the driver to make an informed and well-timed decision whether or not to accept a space- and time-gap. Gap acceptance models have been instrumental in clarifying which factors affect these decisions. These models, developed with support from both naturalistic data and driving simulator studies, typically describe what information drivers use, and predict what decisions they make (Davis & Swenson, 2004; Farah et al., 2009; Hamed et al., 1997; Toledo, 2007). Still, little is known about how the drivers process the relevant perceptual information while arriving to a decision. Understanding and modeling the underlying cognitive mechanisms can lead to more generalizable predictions of these decisions and help to predict how dynamic changes in the environment over the time course of a decision affect the behavior (Jarecki et al.,2020).
Laboratory studies of abstract, highly controlled tasks have pointed towards noisy evidence accumulation as the primary mechanism governing human decision making (Gold & Shadlen, 2007; Ratcliff & McKoon, 2008; Stine et al., 2020). However, it is unclear whether the mechanisms implicated in simple perceptual decisions in the laboratory are as paramount to decisions that are part of complex behaviors like driving. In tasks such as gap acceptance, the available perceptual information is much richer, often varying continuously over time, and motor behaviors are more complex than in traditional laboratory-based decision-making tasks such as motion or orientation discrimination. Even carefully designed abstract tasks can deprive human subjects of the potentially relevant context, the senses of agency and embodiment (Matusz et al., 2019; Shamay-Tsoory & Mendelsohn, 2019). This could hinder understanding of the cognitive mechanisms of interest, and it is far from obvious how to generalize from simple laboratory paradigms on perceptual choice to driving situations. Recently, first steps have been taken toward applying evidence accumulation models in driving and road traffic contexts, but these have been limited to speeded responses to discrete stimuli (Ratcliff, 2015; Ratcliff & Strayer, 2014), or have not been fully stringent in model analyses (Giles et al., 2019; Markkula, Boer, et al., 2018; Xue et al., 2018). Therefore, there is currently a lack of principled investigations of evidence accumulation in contexts where drivers need to consider more complex traffic situations where the correct response is not immediately clear from the stimulus.
The aim of this paper is to bridge the gap between studies of human road user behavior and models of decision making elaborated in basic cognitive science. To this end, we developed an experimental method that allowed drivers to make a decision that is very common in everyday driving—accepting or rejecting a gap when turning across oncoming traffic—but in a way that simultaneously permitted stringent modeling. Using this method, we measured how the participants’ decisions and response times varied with distance and time-to-arrival of the oncoming vehicle. We developed a model to capture the underlying cognitive processes determining the gap acceptance decisions, representing the drivers’ decision making as a dynamic process based on accumulation of noisy evidence over time. We then fitted the developed model to the individual participants’ data. The model captured the between- and within-participant variability in decision outcomes and response times, and successfully predicted the observed behavior in out-of-sample conditions. The results suggest that evidence accumulation of distance and time-to-arrival under time pressure underlies drivers’ decision making during left-turn gap acceptance, and illustrate how simple cognitive models can explain human road user behavior.
METHOD
This research complied with the tenets of the Declaration of Helsinki and was approved by the Human Research Ethics Committee of TU Delft. Informed consent was obtained from each participant.
Participants
Sixteen participants (mean age 27 (range: 22 to 34) years old; 10 male, 6 female) performed a virtual driving task in exchange for a gift voucher worth €15. The participants had been in possession of a driving license for 8 (range: 3 to 18) years on average, of which they have been driving regularly (as interpreted by the participant) for the average of 5 (range: 0 to 18) years.
Setup
The participants performed the experiment in a first-person-view fixed-base driving simulator, which included a 65-inch screen and a commercially available Logitech G29 steering wheel (Figure 1). The distance between the center of the screen and the participants’ eyes was approximately 1.5 m. Carla (Dosovitskiy et al., 2017) was used as a simulation software.
Figure 1.
Experimental setup.
Protocol
Each participant was asked to drive in a virtual urban area (1.5 by 1.5 km, a regular grid of square blocks 150 by 150 m each), following eight randomly generated routes. Each route included 25 intersections; the participants were instructed to go straight on 5 intersections, turn right on 5 intersections, and turn left on 15 intersections (the order of the turns was randomized). This resulted in 120 left turns per participant in total.
The instructions were provided to the participants via auditory navigation prompts which were repeated twice, 120 m and 30 m before the intersection. The participants were told that there will be traffic coming from the opposite direction, and were instructed to fully stop the vehicle at the stopping line on each intersection, which was also marked by the stop sign.
During left turn trials (Figure 2), at the moment the driver stopped, an oncoming vehicle instantaneously appeared across the intersection at a distance of 90, 120, or 150 m (randomly chosen for each trial independently) from the ego vehicle. The initial time-to-arrival (TTA, 4, 5, or 6 s) of the oncoming vehicle was randomly chosen for each left turn. Time-to-arrival conditions were balanced within each route, such that on each route there were exactly five left turns at each of the three time-to-arrival levels. To summarize, we used a 3-by-3 factorial design, varying initial time-to-arrival and distance to the oncoming vehicle.
Figure 2.
Top-down view of the left-turn interaction.
The speed of the oncoming vehicle at the moment of its appearance at the intersection was calculated as the ratio of initial distance to initial time-to-arrival (Table 1), and was held constant at that level throughout the trial. The participants were required to decide whether to go before the oncoming car arrived, or stay, that is, wait until the car passed the intersection and only then continue the route. Positions, velocities, and accelerations of the two vehicles during the interaction were recorded at 100 Hz.
TABLE 1:
Experimental Conditions and the Associated Values of the Speed (m/s) of the Oncoming Vehicle.
| distance∖TTA | 4s | 5s | 6s |
|---|---|---|---|
| 90m | 22.5 | 18 | 15 |
| 120m | 30 | 24 | 20 |
| 150m | 37.5 | 30 | 25 |
Data Analysis
Initial time-to-arrival and distance were the independent variables. Decision (go or stay) and response time (only for “go” decisions) were the dependent variables. The decision was determined based on whether the participant crossed the intersection before the oncoming vehicle. Response time was defined as the time between the appearance of the oncoming vehicle and the time when the participant first pressed the gas pedal after the oncoming vehicle appeared (Figure 3). For “turn” decisions, response time therefore combined the time it took for the driver to make a decision and perceptual and motor delays directly associated with the decision. For “stay” decisions however, appropriate response time measure could not be extracted from the data because unlike “go” decisions, there are no cues in the recorded data that mark the moment when the “stay” decision was made.
Figure 3.
Time traces of a typical interaction, showing the participant decelerating for the stop sign, the oncoming car appearing at a constant speed at t = 0, the dynamics of distance and time-to-arrival (TTA), and the moment when the participant presses the gas pedal to cross in front of the oncoming car.
In this paper, we focus on left-turn gap acceptance decisions. We therefore did not analyze the trials in which the participants did not evaluate a gap, which was the case if they a) performed right turns; b) went straight through the intersection, or c) did not fully stop before the left turn so the oncoming car did not appear at the intersection; the latter happened in 27 (1.4%) left turns. Finally, we excluded 10 (0.5%) “go” decisions in which response time could not be determined because the participant had the gas pedal pressed throughout the trial, and 1 left-turn decision with extreme response time (greater than 2s). This resulted in 1878 analyzed left turns.
Statistical Analysis
Mixed-effects models were used to analyze the effects of distance and TTA on decision outcome (binomial model) and response time (linear model); the models were implemented using R package lme4. In both models, random effects of participant were included to account for individual differences, with the maximum random effects structure permitting model convergence (see the analysis scripts for details). “Stay” decisions were coded as 0, “go” decisions as 1.
Data and Code Availability
Supplementary information, data collection code, analysis scripts, and the data are available at https://osf.io/x3ns6/.
EXPERIMENTAL RESULTS
On average, the participants decided to turn before the arrival of the oncoming car (“go”) in 47% and to wait (“stay”) in 53% of left turns. All trials were included in the decision probability analyses, but only “go” trials were included in the response time analyses.
Replicating existing findings from naturalistic studies of driving (Davis & Swenson, 2004; Patil & Pawar, 2014), we found that probability of a decision to go increased with initial time-to-arrival (z = 9.1, p < 0.001) and initial distance (z = 9.4, p < 0.001) to the oncoming vehicle (Table 2). The existing literature does not report on timing of gap acceptance; interestingly, we found that response time increased with initial time-to-arrival (t = 4.7, p < 0.001), but there was no evidence that it was affected by initial distance (t = −1.4, p = 0.16) (Table 3).
TABLE 2:
Results of Statistical Analysis of the Effect of Distance and TTA Conditions on Decision. The Binomial Mixed-Effects Model Included Random Slopes of Distance and TTA Conditions per Participant.
| Estimate | Std. Error | z value | Pr( z|) | |
|---|---|---|---|---|
| Intercept | −0.28 | 0.28 | −1 | 0.31 |
| TTA condition | 1 | 0.11 | 9.1 | 5.7 × 10−20 |
| Distance condition | 1.4 | 0.15 | 9.4 | 5.2 × 10−21 |
TABLE 3:
Results of Statistical Analysis of the Effect of Distance and TTA Conditions on Response Time in “Go” Decisions. The Linear Mixed-Effects Model Included Random Intercept per Participant.
| Estimate | Std. Error | Df | t value | Pr( t|) | |
|---|---|---|---|---|---|
| Intercept | −0.63 | 0.02 | 16 | −31 | 1.7 × 10−15 |
| TTA condition | 0.023 | 0.0049 | 8.7 × 102 | 4.7 | 3.1 × 10−6 |
| Distance condition | −0.0072 | 0.0052 | 8.7 × 102 | −1.4 | 0.16 |
EVIDENCE ACCUMULATION MODEL OF GAP ACCEPTANCE DECISIONS
We hypothesized that the decision-making process of the driver when choosing whether to stay or go is based on noisy integration of sensory evidence. This integration is subject to noise, and is terminated when sufficient evidence is accumulated. Such decision-making processes are often described by the drift-diffusion model (Bogacz et al., 2006; Ratcliff & McKoon, 2008), which we adopt as a basis for our model. However, the complexity of our gap acceptance task poses fundamental challenges for the classical drift-diffusion model and the likes (see online supplementary information). We address this issue by generalizing the drift-diffusion model in two ways: first, the evidence accumulation rate is determined by a time-varying perceptual evidence, and, second, the amount of evidence that needs to be accumulated decreases with closing time gap (Figure 4).
Figure 4.
Diagram of the drift-diffusion model of driver’s decision making. Relative evidence in favor of going over staying is accumulated until it reaches either a positive or a negative boundary, at which point a “go” or “stay” decision is made, respectively. The rate of accumulation depends on a linear combination of the current values of TTA and d (upper panel). Boundaries for two TTA conditions are plotted for comparison. For each trace, the markers illustrate the moment when the decision is made at the boundaries corresponding to TTA|t=0 = 4s and 6s. For simplicity, non-decision time is illustrated as a perception delay before the start of evidence accumulation; in our model, non-decision time also accounts for the response delay after the decision is made. Time-varying perceptual evidence determines accumulation rate.
First, in our left-turn scenario, the incoming evidence dynamically varies over the time course of each decision. In this scenario, the main perceptual quantities driving the drivers’ decisions are distance and time-to-arrival (Table 2, see also Davis and Swenson (2004)). However, unlike typical laboratory experiments on decision making, these perceptual quantities systematically and substantially change during several hundreds of milliseconds needed to make a decision. Therefore, the canonical assumption of fixed accumulation rate does not hold in our case. To address this, our model hypothesizes that the accumulation rate dynamically depends on distance and time-to-arrival, and is therefore time-dependent.
Assuming that evidence accumulation is driven by a linear combination of distance and time-to-arrival (Giles et al., 2019), its dynamics can be described by the stochastic differential equation
| (1) |
where x is the decision variable, TTA = TTA(t) is the time-to-arrival of the oncoming car at time t, d = d(t) is the distance to the oncoming car at time t, and W is a stochastic Wiener process. The drift rate parameter α ≥ 0 quantifies the relative contribution of incoming perceptual information to the accumulated evidence (decisions are made at random if α = 0). Relative weighting of distance information (compared to time-to-arrival) is characterized by parameter β. Parameter θcrit determines the “critical” value of TTA + βd, such that at the time TTA + βd = θcrit, the drift rate changes sign.
Task Constraints Drive Urgency Effect Via Collapsing Boundaries
Second, with the oncoming car approaching the intersection, the window of opportunity for the driver to go closes fast, which inherently creates an urgency effect: When the time-to-arrival of the oncoming vehicle is large, there is no time pressure for the driver to make a decision, and the driver therefore can sample as much evidence as needed to make an informed decision. However, when the time-to-arrival is small, the driver is compelled to finalize the decision as fast as possible in order to allow enough time for the turning maneuver in case the decision is made to go. Previously, similar time pressure effects in evidence accumulation have been modeled by decision boundaries collapsing over time (Churchland et al., 2008; Drugowitsch et al., 2012). We adopt the same approach in our model, with one important distinction. As the decision urgency in our setup increases with decreasing TTA, we assume the decision boundary to collapse with TTA: the lower the TTA, the less evidence the driver needs to accumulate in order to arrive to a decision.
To capture these considerations, our model assumes that the dynamic accumulation process (1) is terminated when evidence x hits one of the boundaries
| (2) |
where b0 is the boundary scale parameter, and f(⋅) is an increasing function of TTA such that
For simplicity, we chose a sigmoid function satisfying these conditions
| (3) |
where the parameter k > 0 defines the sensitivity of boundary to time-to-arrival, and τ is the time-to-arrival at which the boundary is at its baseline value (±1/2b0).
Perception and Action Delays
Our model only captures the decision process itself, and does not represent sensory perception and decision execution, and, consequently, any time delays associated with them. These delays are necessarily present in the experimentally measured response times, and are typically modeled by singling out a non-decision component of the overall response time (i.e., response time is the sum of decision time and non-decision time) (Ratcliff & McKoon, 2008; Ratcliff & Tuerlinckx, 2002). We included a normally distributed non-decision time in the model
| (4) |
Model Fitting
In total, our model (1)–(4) has eight free parameters: α, β, θcrit, b0, k, τ, μND, σND. We implemented the model in pyddm (Shinn et al., 2020b), and fitted it to the data using differential evolution optimization of the weighted least-sum score (Ratcliff & Tuerlinckx, 2002).
To assess the explanatory power of the model, we fitted it to the data obtained from each participant individually. In addition, to characterize the typical behavior over all participants, we fitted the model to the group-averaged probabilities and response time distributions. To quantify the group-averaged response time distributions across participants, we used the vincentizing approach (Ratcliff, 1979; Rouder & Speckman, 2004; Vincent, 1912): based on the individual participants’ data, we calculated per-participant RT quantile functions, which were then averaged across participants. The group-averaged cumulative distribution function was then calculated as an inverse of the group-averaged quantile function. The group means of response times and probability of going were calculated as the average of within-participant mean values.
MODELING RESULTS
Evidence Accumulation Model Explains and Predicts the Experimental Results
The model explained the observed effect of distance and time-to-arrival on probability of a go decision (Figure 5). Not all participants were described equally well, in particular, the behavior of participant 16 posed a problem for the model. However, overall the model exhibited a full range of behaviors observed in the participants, capturing the behavior of participants whose decisions were relatively weakly affected by the distance condition (e.g., participant 14), as well as those who seemed to decide mostly based on the distance and not time-to-arrival (e.g., participants 10 and 13).
Figure 5.
Probability of “go” decision as a function of distance and time-to-arrival of the oncoming car: Individual participants’ and group-averaged data (markers) and behavior of the model fitted to the respective datasets (lines). The markers in the “all participants” panel represent probabilities averaged over within-participant mean values. Error bars denote binomial proportion standard error of mean.
Crucially, the model captured the observed response times of all sixteen drivers, as well as the group-averaged response times. Specifically, the model explained the positive relationship between response time and time-to-arrival in the group-averaged data (Figure 6). Notably, the model could capture a range of the effect sizes across participants (e.g., participant 1 vs participant 3).
Figure 6.
Response time in the “go” decisions as a function of distance and time-to-arrival of the oncoming car: Individual participants’ and group-averaged data (markers) and behavior of the model fitted to the respective datasets (lines). The markers in the “all participants” panel represents response times averaged over within-participant mean values. Response times for conditions with less than four data points are omitted. Error bars indicate standard error of mean.
Full response time distributions provide further evidence supporting our model. The model fitted to the group response times generated cumulative distribution functions which followed the group-averaged estimates in eight out of nine conditions (Figure 7). In the remaining condition with the shortest TTA and smallest distance (TTA = 4s, d = 90m) however, there were not enough “go” responses to reconstruct the full distribution (only 11 response times across 16 participants). The match between the group-averaged distributions and the model in all other conditions however indicates that the model did not simply fit the “go” probability and mean response times, but also captured the full range of response times generated by the participants.
Figure 7.
Full distributions of response time in “go” decisions: group-averaged cumulative distribution functions (markers) and distributions generated by the model fitted to the group-averaged response times (solid lines). Distribution for TTA = 4s, d = 90m is not visualized because all participants together made only 11 “go” decisions in this condition.
In all analyses above, the model was fitted to all data available from an individual participant, or the group-averaged data. This allowed us to assess whether the model could explain the observed behavior, but not whether it generalizes beyond the data used for fitting, which would not be the case, for instance, if the model was overfitted.
To investigate how the model generalizes to out-of-sample conditions, and at the same time highlight any potential cases of overfitting, we performed cross-validation of the model using a hold-one-condition-out procedure, focusing on the group-averaged data. For each of the nine (TTA, d) conditions, we fitted the model to the data from the remaining eight conditions. The fitted parameters were then used to predict the decision outcomes and mean response times in the ninth, held-out condition. This was repeated nine times, for nine held-out combinations of time and distance gaps.
The model predictions in this cross-validation setting still matched the observed decision probabilities and response time (Figure 8), albeit less well than when fitting to full data. Of all conditions, the response times predicted by the model deviated furthest from the data for TTA = 4s, d = 150m. This indicates a certain degree of overfitting, possibly highlighting that this condition (yielding in a somewhat extreme oncoming car speed of 37.5 m/s) is close to the edge of the model’s feasibility scope. Notably, for other conditions excluding parts of the data from the training set did not substantially affect the simulated model behavior and therefore the match between its predictions and the data. Overall, the cross-validation results suggest that the model’s predictions of the “average” participant’s behavior can generalize to out-of-sample conditions.
Figure 8.
Hold-one-condition-out cross-validation of the model. Filled markers represent the group-averaged data (same as in the “all participants” panels in Figures 5 and 6). Hollow markers indicate model-generated predictions for the held-out condition. Dashed lines connect the hollow markers for visual clarity and do not represent outputs of the model.
Comparison to Simpler Evidence Accumulation Models
To investigate whether our observations could be explained by simpler evidence accumulations models, we have fitted two simplified versions of our model to the data (Figure 9, see also online supplementary information). The first alternative model had the constant drift rate and constant decision bound; the second one included time-varying drift rate but constant decision bound. Both these models could account for the decision outcome data as accurately as the model considered here. However, these models failed to capture the observed effect of time-to-arrival on response time. This suggests that both the time-varying evidence accumulation rate and collapsing decision bounds play an essential role in the participants’ behavior.
Figure 9.
Comparison of the model with simpler alternative models. Alternative model 1: basic drift-diffusion model (constant drift rate, constant decision bound); Alternative model 2: time-varying drift rate, constant decision bound.
DISCUSSION
Human drivers make high-stakes decisions on a daily basis; understanding and predicting these decisions can facilitate the design of safe and efficient transportation systems. To develop such an understanding of left-turn decisions, we build upon the classical drift-diffusion model, and extend it with two key notions. First, our model has a variable drift rate linked to the relevant time-varying perceptual quantities—distance and time-to-arrival to the oncoming vehicle—which are dynamically sampled from the environment. Second, the model relies on collapsing decision boundaries, reflecting the impending constraints imposed on the decision execution by the environment. The model explained the observed decision outcomes and response times, as well as substantial individual differences in both. Cross-validation demonstrated that the model not only explained the data, but also generalized to out-of-sample conditions, effectively providing a way to predict human drivers’ behavior.
Models of gap acceptance have traditionally focused only on the final outcome of a decision, and the factors affecting it (Davis & Swenson, 2004; Farah et al., 2009; Hamed et al., 1997; Toledo, 2007). Our work adds to this literature by emphasizing the process of a decision. In practice, process models of traffic decisions yielding predictions of response times can help to optimize safety and traffic flow efficiency (Markkula, Romano, et al., 2018). More fundamentally, modeling the process of drivers’ decisions provides an insight into the mechanisms of decision making, contributing to theoretical understanding of driver behavior.
Our study is not the first to suggest that evidence accumulation can play a role in behavior on the road; models of this nature have been applied to driver steering and braking decisions during basic sensorimotor control of the vehicle (Markkula, Boer, et al., 2018; Ratcliff, 2015; Ratcliff & Strayer, 2014; Xue et al., 2018), as well as to more complex decisions such as pedestrian road crossing (Giles et al., 2019; Markkula, Romano, et al., 2018). Some of this research has achieved high methodological rigor, on par with the laboratory research on decision making, by collecting large number of data points per participant (Ratcliff, 2015; Ratcliff & Strayer, 2014). To make this possible, these studies used paradigms with speeded response to discrete-onset stimuli, such as brake lights, thus not addressing the types of continuously time-varying sensory information that is crucial in many aspects of driving (Markkula et al., 2016; Toledo, 2007). Other work has considered time-varying sensory information, but has not been as rigorous in the model-fitting approach, for example, absorbing both intra-individual and inter-individual variability into the evidence accumulation noise (Giles et al., 2019; Markkula, Boer, et al., 2018; Xue et al., 2018). The crossing behavior modeling study most similar to ours was that of Giles et al. (2019), who aimed to capture pedestrian crossing decisions. They fitted a number of model candidates to the empirical data, but by own admission they were not able to draw firm conclusions due to limitations in their model-fitting approach. Therefore, to the best of our knowledge, our study is the first to demonstrate methodologically stringent evidence for the value of this type of decision-making model in a driving task with continuously varying sensory evidence.
One of the key notions highlighted by our results is that decisions based on time-varying evidence can be modeled by a drift-diffusion model with the drift rate directly linked to the relevant dynamic perceptual quantities in the environment. This goes beyond the traditional approaches to decision making. These approaches almost ubiquitously focus on the stimuli with the magnitude which is constant throughout the trial, which is naturally modeled by evidence accumulation with constant drift rate (Bogacz et al., 2006; Ratcliff & McKoon, 2008; Ratcliff et al., 2016). In laboratory setting, variable-drift evidence accumulation is rarely investigated, mostly as a way of modeling urgency signals (Cisek et al., 2009; Murphy et al., 2016), non-uniform temporal weighting of evidence throughout the trial (Bronfman et al., 2016; Eckhoff et al., 2008), and artificially introduced perturbances in sensory evidence (Cheadle et al., 2014; Shinn, Ehrlich, et al., 2020). The previous studies on action initiation in intermittent motor control (Markkula, Boer, et al., 2018; Zgonnikov & Markkula, 2018) and the early evidence accumulation based models of traffic interactions discussed above (Giles et al., 2019; Markkula, Romano, et al., 2018) highlighted the promise of the variable-drift approach. However, due to complexity of the analyzed tasks these studies could not disentangle the effects of evidence accumulation and other mechanisms at play. Reinforcing these initial results, our study provides conclusive evidence that evidence accumulation with variable drift can explain gap acceptance decisions made based on naturalistic time-varying stimuli.
Our results also emphasize the crucial role of time pressure induced by the impending oncoming car. The experimental data revealed that response times increased with time-to-arrival in most participants (Figure 6). This goes against predictions of drift-diffusion models with fixed boundaries that imply faster response times for a given decision (in our case, “go”) when perceptual information from the environment (in our case, long TTA) favors that decision. The mechanism of our model which allowed it to capture this experimental observation is decision boundaries collapsing with time-to-arrival. Our scenario is therefore interestingly different from typical speed-accuracy trade-off paradigms in decision-making research, in which time pressures are external to the task being performed (Churchland et al., 2008; Drugowitsch et al., 2012; Shinn, Ehrlich, et al., 2020). In contrast, in the left-turn decisions, the perceptual information being decided upon (time-to-arrival) simultaneously dictates the urgency. Indeed, the time gap which is closing fast imposes an inherent dynamic constraint on decision execution, which the driver has to take into account when deciding.
Our work goes beyond the decision outcomes and begins to uncover the evidence accumulation mechanisms behind the left-turn decisions, but we consider it to be just the first step towards a comprehensive account of drivers’ decision making in gap acceptance. Future work should address the key limitation of our study—the lack of a reliable proxy for the duration of the decision process in “stay” decisions. Traditionally, response times for both options are used to constrain the fits of evidence accumulation models. However, in our experiment, the measurement of “stay” response times could not be extracted from the vehicle motion data. This could potentially be aided, for example, by asking the participants to press a button as soon as they decide to stay. We however opted for not doing so, because such a button press would not be as automatized a behavior as pushing the gas pedal, which could lead to inflated non-decision time in the “stay” decisions and interfere with the decision-making process in “go” decisions. Follow-up studies should develop ecologically valid ways of capturing “stay” response times.
Duration of a decision process provides a useful but coarse-grained measurement of a cognitive process. Going beyond response times, dynamic measures such as movement trajectories can provide researchers with rich data about a cognitive process (Schulte-Mecklenbeck et al., 2017; Zgonnikov et al., 2019). This includes, for instance, information on gradual or abrupt reversal of preference prior to commitment to the final decision (changes-of-mind). Our data indicates that up to 15% of “go” trials could potentially include a change-of-mind when a driver first accelerates soon after seeing the oncoming car, but then stops short of taking the turn, waiting until the car passes instead. However, within our experimental procedure, such changes-of-mind cannot be reliably distinguished from “rolling” behaviors when the drivers simply inch forward after making the “stay” decision in order to take the turn faster after the oncoming vehicle had passed. This is further complicated in case of changes from “stay” to “go,” which cannot be identified based on vehicle motion data alone. Developing a way of reliably identifying changes-of-mind can enable modeling them, which can in turn clarify to what extent they are caused by continued accumulation of new evidence (Fleming et al., 2018; Resulaj et al., 2009) or high level of uncertainty about the initial decision (Atiya et al., 2019; Atiya et al., 2020).
A general limitation of the evidence accumulation models which also extends to our model is the lack of mechanistic explanations of post-decision processes such as decision execution (Evans & Wagenmakers, 2019). In the context of our task, this means that our model cannot account, for example, for the rate of acceleration applied after the “go” decision is made. One possible approach to extend our model in this direction could be through decision confidence (Kepecs et al., 2008; Kiani & Shadlen, 2009): hypothetically, the more confident the driver is in their “go” decision, the stronger they press the gas pedal to execute it. However, this approach requires further experimental investigations to collect per-trial confidence judgments in a non-intrusive manner. Furthermore, it is not clear whether the drift-diffusion model is an adequate framework for capturing decision confidence (Ratcliff et al., 2016; Yeung & Summerfield, 2012). Future studies should investigate the relation between evidence accumulation and response execution in drivers’ decisions both experimentally and computationally.
In addition to theoretical insight, our work has potential practical applications for implementing human models in real-time prediction of human behavior in traffic, as well as virtual environments for testing automated vehicles. Most research on real-time human prediction in traffic uses computationally convenient but cognitively implausible models of human behavior, for instance, fixed decision probability (Kooij et al., 2019; Thornton et al., 2019) and utility maximization (Sadigh et al., 2018; Schwarting et al., 2019), which is known to diverge substantially from actual human behavior (Kahneman & Tversky, 1979), including that in driving (Schmidt-Daffy, 2014; Siebinga et al., 2022). Therefore, having cognitively grounded models of human behavior in traffic interactions can help to generate more accurate and generalizable predictions of road user behavior. Our study provides a step forward here, but needs to be complemented with further work on other scenarios, and translated to real-world contexts with higher variability than in typical controlled studies.
Taken together, our results suggest that dynamic evidence accumulation underlies left-turn gap acceptance decisions in human drivers. More generally, this study exemplifies how simple cognitive process models can help us to understand human behavior in complex real-world tasks.
SUPPLEMENTAL MATERIAL
Supplemental Material for Should I Stay or Should I Go? Cognitive Modeling of Left-Turn Gap Acceptance Decisions in Human Drivers by Arkady Zgonnikov, David Abbink, and Gustav Markkula in Human Factors
ACKNOWLEDGMENTS
We thank Joris Giltay for the help in configuring the driving simulator, and Catholijn Jonker for helpful feedback on the earlier versions of this manuscript. We thank VectorZero for a free license of the RoadRunner software used to design the road layout for the experiment. This research was funded in part by the UK Engineering and Physical Sciences Research Council (EPSRC) grant EP/S005056/1.
Biography
Arkady Zgonnikov is an assistant professor in the Department of Cognitive Robotics at Delft University of Technology, Netherlands. He received his PhD in computer science and engineering from the University of Aizu in 2014.
David A. Abbink is a professor in the Department of Cognitive Robotics at the Delft University of Technology, Netherlands. He received his PhD in mechanical engineering from Delft University of Technology in 2006.
Gustav Markkula is a professor in applied behavior modeling in the Institute for Transport Studies at the University of Leeds, UK. He received his PhD in machine and vehicle systems from Chalmers University of Technology in 2015.
- We measured the time course of drivers’ decision making during left-turn gap acceptance.
- We developed a cognitive process model of drivers’ decisions which incorporates noisy accumulation of dynamic evidence and the urgency effect induced by the task dynamics.
- The model explained the observed decisions and their timing, which suggests that dynamic evidence accumulation is the key cognitive mechanism behind gap acceptance decisions in human drivers.
SUPPLEMENTAL MATERIAL: Supplemental material for this article is available online.
ORCID iD
Arkady Zgonnikov https://orcid.org/0000-0002-6593-6948
David Abbink http://orcid.org/0000-0001-7778-0090
Gustav Markkula http://orcid.org/0000-0003-0244-1582
REFERENCES
- Atiya N. A. A., Rañó I., Prasad G., Wong-Lin K. (2019). A neural circuit model of decision uncertainty and change-of-mind. Nature Communications, 10(1), 2287. 10.1038/s41467-019-10316-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Atiya N. A. A., Zgonnikov A., O’Hora D., Schoemann M., Scherbaum S., Wong-Lin K. (2020). Changes-of-mind in the absence of new post-decision evidence [Publisher: Public Library of Science]. PLOS Computational Biology, 16(2), Article e1007149. 10.1371/journal.pcbi.1007149 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bogacz R., Brown E., Moehlis J., Holmes P., Cohen J. D. (2006). The physics of optimal decision making: A formal analysis of models of performance in two-alternative forced-choice tasks. Psychological Review, 113(4), 700–765. 10.1037/0033-295X.113.4.700 [DOI] [PubMed] [Google Scholar]
- Bronfman Z. Z., Brezis N., Usher M. (2016). Nonmonotonic temporal-weighting indicates a dynamically modulated evidence-integration mechanism [ISBN: 1553-734X Publisher: Public Library of Science San Francisco, CA USA]. PLoS computational biology, 12(2), Article e1004667. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Cheadle S., Wyart V., Tsetsos K., Myers N., De Gardelle V., Castañón S. H., Summerfield C. (2014). Adaptive gain control during human perceptual choice [ISBN: 0896-6273 Publisher: Elsevier]. Neuron, 81(6), 1429–1441. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Churchland A. K., Kiani R., Shadlen M. N. (2008). Decision-making with multiple alternatives. Nature Neuroscience, 11(6), 693–702. 10.1038/nn.2123 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Cisek P., Puskas G. A., El-Murr S. (2009). Decisions in changing conditions: The urgency-gating model [ISBN: 0270-6474 Publisher: Soc Neuroscience]. Journal of Neuroscience, 29(37), 11560–11571. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Davis G. A., Swenson T. (2004). Field study of gap acceptance by left-turning drivers. Transportation Research Record, 1899(1), 71–75. 10.3141/1899-09 [DOI] [Google Scholar]
- Dosovitskiy A., Ros G., Codevilla F., Lopez A., Koltun V. (2017). CARLA: An Open Urban Driving Simulator [ISSN: 2640-3498]. Proceedings of the 1st Annual Conference on Robot Learning, Mountain View, CA, 13-15 November 2017, 1–16. Retrieved October 1, 2022. https://proceedings.mlr.press/v78/dosovitskiy17a.html
- Drugowitsch J., Moreno-Bote R., Churchland A. K., Shadlen M. N., Pouget A. (2012). The cost of accumulating evidence in perceptual decision making. Journal of Neuroscience, 32(11), 3612–3628. 10.1523/JNEUROSCI.4010-11.2012 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Eckhoff P., Holmes P., Law C., Connolly P. M., Gold J. I. (2008). On diffusion processes with variable drift rates as models for decision making during learning. New Journal of Physics, 10(1), 015006. 10.1088/1367-2630/10/1/015006 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Evans N. J., Wagenmakers E.-J. (2019). Evidence Accumulation Models: Current Limitations and Future Directions (tech. rep.) [type: article]. PsyArXiv. 10.31234/osf.io/74df9 [DOI] [Google Scholar]
- Farah H., Bekhor S., Polus A., Toledo T. (2009). A passing gap acceptance model for two-lane rural highways. Transportmetrica, 5(3), 159–172. 10.1080/18128600902721899 [DOI] [Google Scholar]
- Fleming S. M., van der Putten E. J., Daw N. D. (2018). Neural mediators of changes of mind about perceptual decisions. Nature Neuroscience, 21(4), 617–624. 10.1038/s41593-018-0104-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Giles O., Markkula G., Pekkanen J., Yokota N., Matsunaga N., Merat N., Daimon T. (2019). At the zebra crossing: Modelling complex decision processes with variable-drift diffusion models. Proceedings of the 41st Annual Meeting of the Cognitive Science Society, Montreal, Canada, 24366-27372 July 2019, Montreal, Canada. [Google Scholar]
- Gold J. I., Shadlen M. N. (2007). The neural basis of decision making. Annual Review of Neuroscience, 30(1), 535–574. 10.1146/annurev.neuro.29.051605.113038 [DOI] [PubMed] [Google Scholar]
- Hamed M. M., Easa S. M., Batayneh R. R. (1997). Disaggregate gap-acceptance model for unsignalized T-intersections. Journal of Transportation Engineering, 123(1), 36–42. [Google Scholar]
- Jarecki J. B., Tan J. H., Jenny M. A. (2020). A framework for building cognitive process models. Psychonomic Bulletin & Review, 27, 1218–1229. 10.3758/s13423-020-01747-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kahneman D., Tversky A. (1979). Prospect theory: An analysis of decision under risk [Publisher: [Wiley, Econometric Society]]. Econometrica, 47(2), 263–291. 10.2307/1914185 [DOI] [Google Scholar]
- Kepecs A., Uchida N., Zariwala H. A., Mainen Z. F. (2008). Neural correlates, computation and behavioural impact of decision confidence. Nature, 455(7210), 227–231. 10.1038/nature07200 [DOI] [PubMed] [Google Scholar]
- Kiani R., Shadlen M. N. (2009). Representation of confidence associated with a decision by neurons in the parietal cortex. Science, 324(5928), 759–764. 10.1126/science.1169405 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kooij J. F. P., Flohr F., Pool E. A. I., Gavrila D. M. (2019). Context-based path prediction for targets with switching dynamics. International Journal of Computer Vision, 127(3), 239–262. 10.1007/s11263-018-1104-4 [DOI] [Google Scholar]
- Markkula G., Boer E., Romano R., Merat N. (2018. a). Sustained sensorimotor control as intermittent decisions about prediction errors: Computational framework and application to ground vehicle steering. Biological Cybernetics, 112(3), 181–207. 10.1007/s00422-017-0743-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Markkula G., Engström J., Lodin J., Bärgman J., Victor T. (2016). A farewell to brake reaction times? Kinematics-dependent brake response in naturalistic rear-end emergencies. Accident Analysis & Prevention, 95, 209–226. 10.1016/j.aap.2016.07.007 [DOI] [PubMed] [Google Scholar]
- Markkula G., Romano R., Madigan R., Fox C. W., Giles O. T., Merat N. (2018. b). Models of human decision-making as tools for estimating and optimizing impacts of vehicle automation. Transportation Research Record: Journal of the Transportation Research Board, 2672(37), 153–163. 10.1177/0361198118792131 [DOI] [Google Scholar]
- Matusz P. J., Dikker S., Huth A. G., Perrodin C. (2019). Are we ready for real-world neuroscience? Journal of Cognitive Neuroscience, 31(3), 327–338. 10.1162/jocne01276 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Murphy P. R., Boonstra E., Nieuwenhuis S. (2016). Global gain modulation generates time-dependent urgency during perceptual choice in humans [ISBN: 2041-1723 Publisher: Nature Publishing Group]. Nature communications, 7(1), 1–15. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Patil G. R., Pawar D. S. (2014). Temporal and spatial gap acceptance for minor road at uncontrolled intersections in India. Transportation Research Record: Journal of the Transportation Research Board, 2461(1), 129–136. 10.3141/2461-16 [DOI] [Google Scholar]
- Ratcliff R. (1979). Group reaction time distributions and an analysis of distribution statistics. Psychological Bulletin, 86(3), 446–461. 10.1037/0033-2909.86.3.446 [DOI] [PubMed] [Google Scholar]
- Ratcliff R. (2015). Modeling one-choice and two-choice driving tasks. Attention, Perception, & Psychophysics, 77(6), 2134–2144. 10.3758/s13414-015-0911-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ratcliff R., McKoon G. (2008). The diffusion decision model: Theory and data for two-choice decision tasks. Neural Computation, 20(4), 873–922. 10.1162/neco.2008.12-06-420 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ratcliff R., Smith P. L., Brown S. D., McKoon G. (2016). Diffusion decision model: Current issues and history. Trends in Cognitive Sciences, 20(4), 260–281. 10.1016/j.tics.2016.01.007 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ratcliff R., Strayer D. (2014). Modeling simple driving tasks with a one-boundary diffusion model. Psychonomic Bulletin & Review, 21(3), 577–589. 10.3758/s13423-013-0541-x [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ratcliff R., Tuerlinckx F. (2002). Estimating parameters of the diffusion model: Approaches to dealing with contaminant reaction times and parameter variability. Psychonomic Bulletin & Review, 9(3), 438–481. 10.3758/BF03196302 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Resulaj A., Kiani R., Wolpert D. M., Shadlen M. N. (2009). Changes of mind in decision-making. Nature, 461(7261), 263–266. 10.1038/nature08275 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rouder J. N., Speckman P. L. (2004). An evaluation of the vincentizing method of forming group-level response time distributions. Psychonomic Bulletin & Review, 11(3), 419–427. 10.3758/BF03196589 [DOI] [PubMed] [Google Scholar]
- Sadigh D., Landolfi N., Sastry S. S., Seshia S. A., Dragan A. D. (2018). Planning for cars that coordinate with people: Leveraging effects on human actions for planning and active information gathering over human internal state. Autonomous Robots, 42(7), 1405–1426. 10.1007/s10514-018-9746-1 [DOI] [Google Scholar]
- Schmidt-Daffy M. (2014). Prospect balancing theory: Bounded rationality of drivers’ speed choice. Accident; Analysis and Prevention, 63, 49–64. 10.1016/j.aap.2013.10.028 [DOI] [PubMed] [Google Scholar]
- Schulte-Mecklenbeck M., Johnson J. G., Böckenholt U., Goldstein D. G., Russo J. E., Sullivan N. J., Willemsen M. C. (2017). Process-tracing methods in decision making: On growing up in the 70s. Current Directions in Psychological Science, 26(5), 442–450. 10.1177/0963721417708229 [DOI] [Google Scholar]
- Schwarting W., Pierson A., Alonso-Mora J., Karaman S., Rus D. (2019). Social behavior for autonomous vehicles. Proceedings of the National Academy of Sciences, 116(50), 201820676. 10.1073/pnas.1820676116 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Shamay-Tsoory S. G., Mendelsohn A. (2019). Real-Life neuroscience: An ecological approach to brain and behavior research: [Publisher: SAGE PublicationsSage CA: Los Angeles, CA]. Perspectives on Psychological Science, 14(5), 841–859. 10.1177/1745691619856350 [DOI] [PubMed] [Google Scholar]
- Shinn M., Ehrlich D. B., Lee D., Murray J. D., Seo H. (2020. a). Confluence of timing and reward biases in perceptual decision-making dynamics [Publisher: Soc Neuroscience]. Journal of Neuroscience, 40(38), 7326–7342. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Shinn M., Lam N. H., Murray J. D. (2020. b). A flexible framework for simulating and fitting generalized drift-diffusion models. eLife, 9, Article e56938. 10.7554/eLife.56938 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Siebinga O., Zgonnikov A., Abbink D. (2022). A human factors approach to validating driver models for interaction-aware automated vehicles. ACM Transactions on Human-Robot Interaction, 11(4), 47:1–47:21. 10.1145/3538705 [DOI] [Google Scholar]
- Stine G. M., Zylberberg A., Ditterich J., Shadlen M. N. (2020).Differentiating between integration and non-integration strategies in perceptual decision making (Wyart V., Frank M. J., Wyart V., Usher M., Eds.) [Publisher: eLife Sciences Publications, Ltd]. eLife, 9, Article e55365. 10.7554/eLife.55365 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Thornton S. M., Limonchik B., Lewis F. E., Kochenderfer M. J., Gerdes J. C. (2019). Toward closing the loop on human values. IEEE Transactions on Intelligent Vehicles, 4(3), 437–446. 10.1109/TIV.2019.2919471 [DOI] [Google Scholar]
- Toledo T. (2007). Driving behaviour: Models and challenges [Publisher: Routledge eprint: https://doi.org/10.1080/01441640600823940]. Transport Reviews, 27(1), 65–84. 10.1080/01441640600823940 [DOI] [Google Scholar]
- Vincent S. B. (1912). The functions of the vibrissae in the behavior of the white rat. Behavior Monographs, 1(5), 7–81. [Google Scholar]
- Xue Q., Markkula G., Yan X., Merat N. (2018). Using perceptual cues for brake response to a lead vehicle: Comparing threshold and accumulator models of visual looming. Accident Analysis & Prevention, 118, 114–124. 10.1016/j.aap.2018.06.006 [DOI] [PubMed] [Google Scholar]
- Yeung N., Summerfield C. (2012). Metacognition in human decision-making: Confidence and error monitoring. Philosophical Transactions of the Royal Society B: Biological Sciences, 367(1594), 1310–1321. 10.1098/rstb.2011.0416 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zgonnikov A., Atiya N. A. A., O’Hora D., Rañó I., Wong-Lin K. (2019). Beyond reach: Do symmetric changes in motor costs affect decision making? A registered report. Judgment and Decision Making, 14(4), 455–469. [Google Scholar]
- Zgonnikov A., Markkula G. (2018). Evidence accumulation account of human operators’ decisions in intermittent control during inverted pendulum balancing. Proceedings of IEEE International Conference on Systems, Man, and Cybernetics (SMC), Miyazaki, Japan, 7716-10721 October 2018, Miyazaki, Japan. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Supplemental Material for Should I Stay or Should I Go? Cognitive Modeling of Left-Turn Gap Acceptance Decisions in Human Drivers by Arkady Zgonnikov, David Abbink, and Gustav Markkula in Human Factors
Data Availability Statement
Supplementary information, data collection code, analysis scripts, and the data are available at https://osf.io/x3ns6/.









