Abstract
Classical models of perceptual decision-making assume that subjects use a single, consistent strategy to form decisions, or that strategies evolve slowly over time. Here we present new analyses suggesting that this common view is incorrect. We analyzed data from mouse and human decision-making experiments and found that choice behavior relies on an interplay between multiple interleaved strategies. These strategies, characterized by states in a hidden Markov model, persist for tens to hundreds of trials before switching, and may alternate multiple times within a session. The identified mouse decision-making strategies were highly consistent across individuals and comprised a single “engaged” state, in which decisions relied heavily on the sensory stimulus, and several biased states in which errors frequently occurred. These results provide a powerful alternate explanation for “lapses” often observed in rodent psychophysical experiments, and suggest that standard measures of performance mask the presence of dramatic changes in strategy across trials.
1. Introduction
Understanding the computations performed in the brain will require a comprehensive characterization of behavior [1–3]. This realization has fueled a recent surge in methods devoted to the measurement, quantification, and modeling of natural behaviors [4–8]. Historically, studies of perceptual decision-making behavior have tended to rely on models derived from signal detection theory (SDT) [9, 10] or evidence accumulation [11–13]. More recently, approaches based on reinforcement learning have also been used to model the effects of context, reward, and trial-history on perceptual decision-making behavior [14–18]. In all cases, however, these approaches describe decision-making in terms of a single strategy that does not change abruptly across trials or sessions.
One puzzling aspect of sensory decision-making behavior is the presence of so-called “lapses”, in which an observer makes an error despite the availability of strong sensory evidence. The term itself suggests an error that arises from a momentary lapse in attention or memory, as opposed to an inability to perceive the sensory stimulus. Lapses arise in all species, but are surprisingly frequent in rodent experiments, where lapses can comprise up to 10–20% of all trials [19–21].
The standard approach for modeling lapses involves augmenting the classic psychometric curve with a “lapse parameter”, which characterizes the probability that the observer simply ignores the stimulus on any given trial [22–24]. This model can be conceived as a mixture model [25] in which, on every trial, the animal flips a biased coin to determine whether or not to pay attention to the stimulus when making its choice. Previous literature has offered a variety of explanations for lapses, including inattention, motor error, and incomplete knowledge of the task [22, 24, 26], and recent work has argued that they reflect an active process of uncertainty-guided exploration [17]. However, a common thread to these explanations is that lapses arise independently across trials, in a way that does not depend on the time course of other lapses.
Here we show that lapses do not arise independently, but depend heavily on latent states that underlie decision-making behavior. We use a modeling framework based on hidden Markov models (HMMs) to show that mice rely on discrete decision-making strategies that persist for tens to hundreds of trials. While the classic lapse model corresponds to a special case in our framework, the model that best described the empirical choice data of real mice had one state corresponding to an “engaged” strategy, in which the animal’s choices were strongly influenced by the sensory stimulus, and other states that corresponded to biased or weakly stimulus-dependent strategies. Our analyses show that lapses arise primarily during long sequences of trials when the animal is in a biased or disengaged state. Conversely, we found that animals with high apparent lapse rates may nevertheless be capable of high-accuracy performance for extended blocks of trials. We applied our modeling framework to datasets from two different mouse decision-making experiments [19, 20], and to one dataset from a decision-making experiment in humans [27]. Taken together, these results shed substantial new light on the factors governing sensory decision-making, and provide a powerful set of tools for identifying hidden states in behavioral data.
2. Results
2.1. The classic lapse model for sensory decision-making
A common approach for analyzing data from two-choice perceptual decision-making experiments involves the psychometric curve, which describes the probability that the animal chooses one option (e.g., “rightward”) as a function of the stimulus value [22–24]. The psychometric curve is commonly parameterized as a sigmoidal function that depends on a linear function of the stimulus plus an offset or bias. This sigmoid rises from a minimum value of γr to a maximal value of 1 − γl, where γr and γl denote “lapse” parameters, which describe the probability of of making a rightward or leftward choice independent of the stimulus value. Thus, the probability of a rightward choice is always at least γr, and it cannot be greater than 1 − γl. In what follows, we will refer to this as the “classic lapse model of choice behavior”, which can be written:
(1) |
where yt ∈ {0, 1} represents the choice (left or right) that an animal makes at trial t, is a vector of covariates, and is a vector of weights that describes how much each covariate influences the animal’s choice. Note that xt includes both the stimulus and a constant ‘1’ element to capture the bias or offset, but it may also include other covariates that empirically influence choice, such as previous choices, stimuli, and rewards [28–30].
Although the classic lapse model can be viewed as defining a particular sigmoid-shaped curve relating the stimulus strength to behavior (Fig. 1c), it can equally be viewed as a mixture model [25]. In this interpretation, we regard the animal as having an internal state zt that takes on one of two different values on each trial, namely “engaged” or “lapse”. If the animal is engaged, it makes its choice according to the classic sigmoid curve (which saturates at 0 and 1). If lapsing, it ignores the stimulus and makes its choice based only on the relative probabilities of a left and right lapse. Mathematically, this can be written:
(2) |
where p(zt = “lapse”) = (γr + γl) and p(zt = “engaged”) = 1 − (γr + γl). In this interpretation, the animal flips a biased coin, with fixed probability (γr + γl), on each trial and then adopts one of two strategies based on the outcome—a strategy that depends on the stimulus vs. one that ignores it. Note that the animal can make a correct choice in the lapse state through pure luck; we use “lapse” here simply to indicate that the animal is not relying on the stimulus when making its decision.
Viewing the classic lapse model as a mixture model highlights some of its limitations. First, it assumes that animals switch between only two decision-making strategies. Second, it assumes that lapses occur independently in time, according to an independent Bernoulli random variable on each trial. Finally, the model assumes that choices in the “lapse” state are fully independent of the stimulus, neglecting the possibility that they are still weakly stimulus dependent [24], or are influenced by other covariates such as reward or choice history [28, 29]. These limitations motivate us to consider a more general family of models, which includes the classic lapse model as a special case.
2.2. A model for decision-making with multiple strategies
Recognizing the limitations of the classic lapse model, we propose to analyze perceptual decision-making behavior using a framework based on Hidden Markov models (HMMs) with Bernoulli generalized linear model (GLM) observations [6, 31]. The resulting “GLM-HMM” framework, also known as an input-output HMM [32], allows for an arbitrary number of states, which can persist for an extended number of trials and exhibit different dependencies on the stimulus and other covariates.
A GLM-HMM has two basic pieces: an HMM governing the distribution over latent states, and a set of state-specific GLMs, specifying the decision-making strategy employed in each state (see Fig. 1). For a GLM-HMM with K latent states, the HMM has a K × K transition matrix A specifying the probability of transitioning from any state to any other,
(3) |
where zt−1 and zt indicate the latent state at trials t − 1 and t, respectively. The “Markov” property of the HMM is that the state on any trial depends only on the state from the previous trial, and the “hidden” property refers to the fact that states are latent or hidden from external observers. For completeness, the HMM also has a distribution over initial states, given by a K-element vector π whose elements sum to 1, giving p(z1 = k) = πk.
To describe the state-dependent mapping from inputs to decisions, the GLM-HMM contains K independent Bernoulli GLMs, each defined by a weight vector specifying how inputs are integrated in that particular state. The probability of a rightward choice (yt = 1) given the input vector xt and the latent state zt is given by
(4) |
where denotes the GLM weights for latent state k ∈ {1, …, K}. The full set of parameters for a GLM-HMM, , are learned directly from each animal’s choice data via the EM algorithm described in Section 4.1.
It is worth noting that the classic lapse model described in Eq. 1 and Eq. 2 corresponds to a restricted 2-state GLM-HMM. If we consider state 1 to be “engaged” and state 2 to be the “lapse” state, then the state-1 GLM has weights w1 = w, and the state-2 GLM has all weights set to 0 except the bias weight, which is equal to −log(γl/γr). The transition matrix has identical rows, with probability 1 − (γr + γl) of going into state 1 and probability (γr + γl) of going into state 2 at the next trial, regardless of the current state. This ensures that the probability of a lapse on any given trial is stimulus-independent and does not depend on the previous trial’s state. Fig. 1a–c shows an illustration of the classic lapse model formulated as a 2-state GLM-HMM.
However, there is no general reason to limit our analyses to this restricted form of the GLM-HMM. By allowing the model to have more than two states, multiple states with non-zero stimulus weights, and transition probabilities that depend on the current state, we obtain a model family with a far richer set of dynamic decision-making behaviors. Fig. 1d shows an example GLM-HMM with three latent states, all of which have high probability of persisting for multiple trials. Intriguingly, the psychometric curve arising from this model (Fig. 1f) is indistinguishable from that of the classic lapse model. Thus, multiple generative processes can result in identical psychometric curves, and we must look beyond the psychometric curve if we want to gain insight into the dynamics of decision-making across trials.
2.3. Mice switch between multiple strategies.
To examine whether animals employ multiple strategies during decision-making, we fit the GLM-HMM to behavioral data from two binary perceptual decision-making tasks (see Methods 4.1). First, we fit the GLM-HMM to choice data from 37 mice performing a visual detection decision-making task developed in [33] and adopted by the International Brain Laboratory (IBL) [19]. During the task, a sinusoidal grating with contrast between 0 and 100% appeared either on the left or right side of the screen (Fig. 2a). The mouse had to indicate this side by turning a wheel. If the mouse turned the wheel in the correct direction, it received a water reward; if incorrect, it received a noise burst and an additional 1-second timeout. During the first 90 trials of each session, the stimulus appeared randomly on the left or right side of the screen with probability 0.5. Subsequent trials were generated in blocks in which the stimulus appeared on one side with probability 0.8, alternating randomly every 20–100 trials. We analyzed data from animals with at least 3000 trials of data (across multiple sessions) after they had successfully learned the task (see Fig. S1, Fig. S2). For each animal, we considered only the data from the first 90 trials of each session, when the stimulus was equally likely to appear on the left or right of the screen.
We modeled the animals’ decision-making strategies using a GLM-HMM with four inputs: (1) the (signed) stimulus contrast, where positive values indicate a right-side grating and negative values indicate a left-side grating; (2) a constant offset or bias; (3) the animal’s choice on the previous trial; and (4) the stimulus side on the previous trial. A large weight on the animal’s previous choice gives rise to a strategy known as “perserveration” in which the animal makes the same choice many times in a row, regardless of whether it receives a reward. A large weight on the previous stimulus side, which we refer to as the “win-stay, lose-switch” regressor, gives rise to the well-known strategy in which the animal repeats a choice if it was rewarded, and switches choices if it was not. Note that for the IBL task in question, bias and trial-history dependencies were sub-optimal, meaning that the maximal-reward strategy was to have a large weight on the stimulus and zero weights on the other three inputs.
To determine the number of different strategies underlying decision-making behavior, we fit GLM-HMMs with varying numbers of latent states. Note that the 1-state model is simply a standard Bernoulli GLM, while the classic lapse model (Eq. 1, Eq. 2) is the constrained 2-state model mentioned previously. We found that a 3-state GLM-HMM substantially outperformed models with fewer states, including the classic lapse model. The states of the fitted model were readily interpretable, and tended to persist for many trials in a row.
Figures 2 and 3 show results for an example mouse. For this animal, the multi-state GLM-HMM outperformed both the standard (1-state) GLM and the classic lapse model, both in test log-likelihood and percent correct, with the improvement approximately leveling off at 3 latent states (Fig. 2b–c). Because the test set for this mouse contained 900 trials, test set log-likelihood increases of 0.13 bits/trial and 0.09 bits/trial for the 3-state model over the 1-state model and classic lapse models, respectively, meant that the data were (20.13)900 ≈ 1.7 × 1035 and (20.09)900 ≈ 2.4 × 1024 times more likely under the 3-state model. Note that while the number of model parameters increased with the number of states in the GLM-HMM (and a two state GLM-HMM has more parameters than both the classic lapse model or single state GLM), the log-likelihood of the test dataset was not guaranteed to increase with more states [34]. Indeed, as the number of states increased, the GLM-HMM could have started to overfit the training dataset causing it to fit the test dataset poorly (see Fig. ED1).
The transition matrix for the fitted 3-state model describes the transition probabilities between three different states, each of which corresponds to a different decision-making strategy (Fig. 2d). Large entries along the diagonal of this matrix, ranging between 0.94 and 0.98, indicate a high probability of remaining in the same state for multiple trials. The other set of inferred parameters were the GLM weights, which define how the animal makes decisions in each state (Fig. 2e). One of these GLMs (“state 1”) had a large weight on the stimulus and negligible weights on other inputs, giving rise to high-accuracy performance on the task (Fig. 2f). The other two GLMs (“state 2” and “state 3”), by comparison, had small smaller weights on the stimulus, and relatively large bias weights.
We can visualize the decision-making strategies associated with these states by plotting the corresponding psychometric curves (Fig. 2g), which show the probability of a rightward choice as a function of the stimulus, conditioned on both previous choice and reward. The steep curve observed in state 1, which corresponds to the mouse achieving near-perfect performance on high-contrast stimuli, led us to adopt the name ‘engaged’ to describe this state. By comparison, the psychometric curves for states 2 and 3 reflected large leftward and rightward biases, respectively. They also had relatively large dependence on previous choice and reward, as indicated by the gap between solid and dashed lines. While this mouse had an overall accuracy of 80%, it achieved 90% accuracy in the engaged state, compared to only 60% and 58% accuracy in the two biased states (Fig. 2f).
To gain insight into the temporal structure of decision-making behavior, we used the fitted 3-state model to compute the posterior probability over the mouse’s latent state across all trials (Fig. 3). The resulting state trajectories reflect our posterior beliefs about the animal’s internal state on every trial, given the entire sequence of observed inputs and choices during a session (See Methods, Sec. 4.1.3).
Contrary to the predictions of the classic lapse model, states tended to persist for many trials in a row. Remarkably, we found that the most probable state often had probability close to 1, indicating a high degree of confidence about the mouse’s internal state given the observed data.
To quantify state occupancies, we assigned each trial to its most probable state, and found that this example mouse spent approximately 69% of all trials in the engaged state (out of 5040 total trials over 56 sessions), compared to 15% and 16% of trials in the biased-left and biased-rightward states (Fig. 3c). Moreover, the mouse changed state at least once within a session in roughly 71% of all 90-trial sessions, and changed multiple times in 59% of sessions (Fig. 3d). This rules out the possibility that the states merely reflect the use of different strategies on different days. Rather, the mouse tended to remain in an engaged, high-performance state for tens of trials at a time, with lapses arising predominantly during interludes when it adopted a left-biased or right-biased strategy for multiple trials in a row. The multi-state GLM-HMM thus provides a very different portrait of mouse decision-making behavior than the basic GLM or lapse model.
2.4. State-based strategies are consistent across mice
To assess the generality of these findings, we fit the GLM-HMM to the choice data from 37 mice in the IBL dataset [19] (Fig. 4). We found that the results shown for the example animal considered above were broadly consistent across animals. Specifically, we found that the 3-state GLM-HMM strongly outperformed the basic GLM and classic lapse model in cross-validation for all 37 mice (Fig. 4a and Fig. S3). On average, it predicted mouse choices with 4.2% higher accuracy than the basic GLM (which had an average prediction accuracy of 78%), and 2.8% higher accuracy than the classic lapse model (Fig. 4b). Furthermore, for one animal the improvement in prediction accuracy for the 3 state GLM-HMM was as high as 12% relative to the basic GLM, and 7% relative to the classic lapse model.
Although performance continued to improve slightly with four and even five latent states, we will focus our analyses on the 3-state model for reasons of simplicity and interpretability. Fig. S4 provides a full analysis of 4-state model, showing that it tended to divide the engaged state from the 3-state model into two “sub-states” that differed slightly in bias.
Fits of the 3-state GLM-HMM exhibited remarkable consistency across mice, with a majority exhibiting states that could be classified as “engaged”, “biased-left”, and “biased-right” (Fig. 4d). (See Methods section 4.1.5 for details about the alignment of states across mice). While we plot inferred transition matrices for all 37 mice in Fig. S5, here we used the diagonal elements of each matrix to compute an expected dwell time for each animal in each state (Fig. 4e). This revealed a median dwell time, across animals, of 24 trials for the engaged state, versus 13 and 12 trials for the biased-left and biased-right states, respectively. This description of behavior departs dramatically from the assumptions of the classic lapse model: for the classic lapse model and a lapse rate of 20%, the expected dwell time in the lapse state is just 1/(1–0.2) = 1.25 trials. We analyzed the distribution of state dwell times inferred from data and found they were well approximated by a geometric distribution, matching the theoretical distribution of data sampled from a Hidden Markov Model (Fig. ED2 and Fig. S6).
We also examined the fraction of trials per session that mice spent in each of the three states (Fig. 4c). To do so, we used the fitted model parameters to compute the posterior probabilities over state, and assigned each trial to its most likely state. The resulting “fractional occupancies” revealed that the median mouse spent 69% of its time in the engaged state, with the best mice exceeding 90% engagement. Moreover, the majority of sessions (83% of 2017 sessions across all animals) involved a switch between two or more states; in only 17% of sessions did a mouse remain in the same state for an entire 90-trial session.
To further quantify the model’s ability to capture temporal structure of choice data, we examined choice run-length statistics, where runs were defined as sequences of trials in which the mouse repeated the same choice. We found that the IBL mice exhibited longer runs than if they performed the task perfectly, and the fitted GLM-HMM reproduced this feature more accurately than the classic model (Fig. ED8).
Finally, to make sure that the states identified by the model were not a consequence of the IBL mice having been previously exposed to blocks of trials with a consistent side bias (see Section 4.4), we examined data from 4 mice that were never exposed to “bias blocks”. In Fig. ED3, we show that the retrieved states, dwell times and model comparison results for these animals were similar to those shown in the rest of the IBL population (shown in Fig. 4).
2.5. Data provide evidence for discrete, not continuous, states
The GLM-HMM describes perceptual decision-making in terms of discrete states that persist for many trials in a row before switching. However, the model’s dramatic improvement over classic models does not guarantee that the states underlying decision-making are best described as discrete. One could imagine, for example, that a state governing the animal’s degree of engagement drifts gradually over time, and that the GLM-HMM simply divides these continuous changes into discrete clusters. To address this possibility, we fit the data with PsyTrack, a psychophysical model with continuous latent states [35, 36]. The PsyTrack model describes sensory decision-making using an identical Bernoulli GLM, but with dynamic weights that drift according to a Gaussian random walk (see Methods sec. 4.3).
For all 37 mice in our dataset, the 3-state GLM-HMM achieved substantially higher test log-likelihood than the PsyTrack model (left panel of Fig. 4f). Model selection also correctly identified simulated data from the GLM-HMM, whereas datasets simulated from a matched first-order autoregressive model had roughly equal log-likelihood under the two models (middle and right panels of Fig. 4f). Not only was the GLM-HMM favored over PsyTrack for IBL mice, but we show that it was favored over PsyTrack for the other mouse datasets we study in this work: see Fig. ED4. Finally, to verify that the GLM-HMM’s improved performance compared to PsyTrack was not merely due to the fact that the GLM-HMM does better at accounting for lapses, we also show the average fits of both PsyTrack and the 3 state GLM-HMM to the empirical choice data of the example mouse in Fig. S7. Despite the absence of explicit lapse parameters in both of these models, both are able to capture the non-zero error rates of the mouse on easy trials.
2.6. Mice switch between multiple strategies in a second task
To ensure that our findings were not specific to the IBL task or training protocol, we examined a second mouse dataset with a different sensory decision-making task. Odoemene et al. [20] trained mice to report whether the flash rate of an LED was above or below 12Hz by making a right or left nose poke (Fig. 5a). Once again, we found that the multi-state GLM-HMM provided a far more accurate description of mouse decision-making than a basic GLM or the classic lapse model (Fig. 5b). Although the performance of the 3-state and 4-state models was similar, we focused on the 4-state model because—in addition to having slightly higher test log-likelihood for a majority of animals (see Fig. S8 for individual curves)—the 4-state model balanced simplicity and interpretability, with each state in the 4-state model corresponding to a distinct behavioral strategy. (See Fig. S9 and Fig. S10 for a comparison to 3-state and 5-state fits). The 4-state model exhibited an average improvement of 0.025 bits/trial over the classic lapse model, making the test dataset approximately 1 × 1018 times more likely under the GLM-HMM than to the lapse model.
Fig. 5d shows the inferred GLM weights associated with each state in the 4-state GLM-HMM. Based on the qualitative differences in the state-specific psychometric curves (Fig. 5e), we labeled the four states as “engaged”, “biased left”, “biased right” and “win-stay”. The combination of stimulus and choice history weights for this fourth state gave rise to a large separation between psychometric curves conditioned on previous reward; the resulting strategy could be described as “win-stay” because the animal tended to repeat a choice if it was rewarded. (However, it did not tend to switch if a choice was unrewarded). Accuracy was highest in the engaged state (92%) and lowest in the biased-left (67%) and biased right states (77%). Similar to the IBL dataset, the identified states tended to persist for many trials in a row before switching (Fig. 5f and Fig. ED5b).
Finally, we used the fitted model to examine the temporal evolution of latent states within a session. Figure 5g shows the average posterior state probabilities over the first 200 trials in a session for two example mice and the average over all mice (Fig. ED6 shows average posterior state probabilities for each individual mouse in the cohort separately). These trajectories reveal that mice typically began a session in one of the two biased states, and had a low probability of entering the engaged state within the first 50 trials: mice used these initial trials of a session to “warm-up” and then gradually improved their performance [14]. This represents a substantial departure from the IBL mice, most of which had a high probability of engagement from the very first trial, and had relatively flat average trajectories over the first 90 trials of a session (Fig. 3 and Fig. S12). Note however that the warm-up effect was not the only form of state switching observed in the Odoemene et al mice, as the animals switched states multiple times per session and continued switching beyond the first 100 trials (Fig. ED5). We also examined whether the effects of fatigue or satiety could be observed in the average state probabilities at the end of sessions, but did not find consistent patterns across animals (Fig. ED7).
2.7. External correlates of engaged and disengaged states
One powerful feature of the GLM-HMM is the fact that it can be used to identify internal states from binary decision-making data alone. However, it is natural to ask whether these states manifest themselves in other observable aspects of mouse behavior. In other words, does a mouse behave differently when it is in the engaged state than in a biased state, above and beyond its increased probability of making a correct choice? To address this question, we examined response times and violation rates, two observable features that we have did not incorporate into the models.
Previous literature has revealed that humans and monkeys commit more errors on long duration trials than short duration trials [37–39]. This motivated us to examine the distributions of response times for the engaged state, compared to for the disengaged (left and right biased) states.
In the IBL dataset, we found that engaged and disengaged response time distributions were statistically different for all 37 mice (Komogorov-Smirnov tests reject the null hypothesis with p < 0.05 for all 37 mice; Fig. 6a). Examining the Q-Q plots, it was also clear that the most extreme response times for each animal were associated with the disengaged states (Fig. 6b). In the Odoemene et al. [20] dataset, we examined differences in the rate of violations—trials where the mouse failed to make a decision within the response period.The mean violation rate across all mice was 21% (much higher than the 1% violation rate observed in the IBL dataset), and violations were 3.2% more common in the disengaged states (2, 3, and 4) than in the engaged state (state 1).
2.8. State dependent decision-making in human visual task
Finally, while the primary focus of this paper is identifying the discrete latent strategies underpinning mouse decision-making, the GLM-HMM framework is general and can be used to study choice behavior across species. We, thus, applied the GLM-HMM to the choice behavior of humans performing the visual motion discrimination task of [27]. Participants had to judge the difference in motion coherence between two consecutive random dot kinematograms: the first kinetogram always corresponded to a reference stimulus with 70% motion coherence, while the second stimulus could have greater or less motion coherence (Figure 7a). Participants indicated the stimulus they perceived to have the greater motion coherence by pressing a button. Cross-validation revealed that the choice data of 24 out of 27 humans was better explained by a 2 or 3 state GLM-HMM compared to the classic lapse model (Fig. 7b–c). The mean improvement of the 2 state GLM-HMM compared to the classic lapse model was 0.013 bits/trial, making a test dataset of 500 trials 90 times more likely to have been generated by a 2 state GLM-HMM than by the classic lapse model.
Fig. 7d shows the retrieved weights for each participant for the 2 state model (we focus on the 2 state model as this was the model favored by the majority of participants), while Fig. 7e shows the probability of the participant selecting, in each of the two states (blue and green curves), that the second stimulus had greater motion coherence than the first (choice = ‘more’) as a function of the relative motion coherence. The 2 states correspond to participants having a bias for selecting each of the two possible outcomes - when using the blue state, participants were biased toward selecting ‘more’, while participants were biased toward selecting ‘less’ in the green state. Fig. 7f shows the expected dwell times for all participants in each state, while Fig. 7g shows the empirical number of state changes detected in each session (across the entire participant population). The median number of state changes per session was 6 switches, indicating that it is not just mice that switch between strategies multiple times per session. Fig. 7h illustrates the strategy changes across a session and shows the posterior state probabilities for 3 example sessions (for 3 different participants).
3. Discussion
In this work, we used the GLM-HMM framework to identify hidden states from perceptual decision-making data. Mouse behavior in two different perceptual decision-making tasks [19, 20] was far better described by a GLM-HMM with sustained engaged and biased states. Unlike the classic lapse model, these states alternated on the timescale of tens to hundreds of trials. We also applied the model to human psychophysical data [27] and found that human behavior was also better described by a GLM-HMM with sustained states that differed in bias. While it may seem obvious that mice should exhibit fluctuations in task engagement over time, this is a very different account of behavior to that given by the classic lapse model, which assumes that error trials do not occur in blocks, and are instead interspersed throughout a session. Furthermore, our method offers the practical benefit that it allows practitioners to segment trials according to the inferred behavioral strategy. Finally, it was not obvious that this alteration in choice policy over the course of a session should occur in a discrete rather than continuous manner. Comparing our results with those for PsyTrack—a model with continuous decision-making states—provided support for the view that mice use discrete rather than continuous decision-making states.
While we found similarities in the strategies pursued by mice performing the different visual detection tasks, we also found some differences. For example, we found that mice performing the Odoemene et al. [20] task used an additional ‘win-stay’ strategy that was not observed in the IBL mice. One possible explanation for these strategy differences is that they arise due to the different shaping protocols associated with the different tasks. A follow-up study could investigate the role of shaping on GLM-HMM behavioral strategies.
The ability to infer the state or strategy employed by an animal on a trial-by-trial basis will be useful for characterizing differences in performance across sessions and across animals. It will also provide a powerful tool for neuroscientists studying the neural mechanisms that support decision-making [40–46]. It may be the case that different strategies rely on different circuits or different patterns of neural activity [47].
Although we found evidence of warm-up behavior at the start of sessions in the Odoemene et al. [20] mice, we were somewhat surprised to find few signatures of satiation or fatigue toward the end of a session. This might be due to the fact that sessions were typically of fixed duration, and may have ended before mice had a chance to grow satiated or fatigued. One future direction will be to apply the GLM-HMM to experiments with longer sessions, where it may be useful for detecting changes in behavior reflecting satiety or fatigue.
Compared to the mice we studied, humans used two relatively engaged strategies with opposite biases when performing the perceptual decision-making task of [27]. Unfortunately, it is impossible to say if the difference in strategies observed across species were due to species differences, or to differences in task protocol. An interesting follow-up experiment would be to apply the GLM-HMM to data from different species performing the same task.
In future work, we will aim to make explicit the connection between the ‘engaged’ and ‘disengaged’ strategies identified by our model and existing measures of arousal and engagement in the literature. Identifying the relationship between the GLM-HMM’s hidden states and pupil diameter, low-frequency LFP oscillations, spontaneous neuronal firing, noise correlations and the action of neuromodulators [48–51] will be a priority.
That discrete states underpin mouse and human choice behavior may also call for new normative models to explain why subjects develop these states to begin with [52]. The existence of disengaged and engaged states could reflect explore-exploit behavior [17, 53, 54], optimal learning (e.g., [55–58]), or could simply indicate incomplete learning of the task. Alternatively, if the states uncovered by our model reflect a similar phenomenon to the internal and external modes of sensory processing observed in [59, 60], it may be that they arise due to the limited information-processing capacity of the brain.
Another promising direction will be to replace the model’s fixed transition matrix with a generalized linear model so as to allow external covariates to modulate the probability of state changes[6]. This would allow us to identify the factors that influence state changes (e.g., a preponderance of unrewarded choices) and, potentially, seek to control such transitions.
While there are many avenues for future research, we believe that these results call for a significant rethinking of rodent and human perceptual decision-making behavior and the methods for analyzing it. Indeed, standard analysis methods do not take account of the possibility that an animal makes abrupt changes in decision-making strategy multiple times per session. We feel that the ability to infer internal states from choice behavior will open up new directions for data analysis and provide new insights into a previously inaccessible dimension of perceptual decision-making behavior.
4. Methods
We confirm that our research complies with all relevant ethical regulations.
4.1. Inference of GLM-HMM parameters
4.1.1. GLM-HMM objective function
When fitting the GLM-HMM, the goal is to learn the transition matrix , the set of weights influencing choice in each state, with , and the initial state distribution, . We fit this set of parameters, collectively labeled as Θ, to choice data using Maximum A Posteriori (MAP) estimation. This was practically implemented by the Expectation Maximization (EM) algorithm [61].
The EM algorithm has been adapted to fit Hidden Markov models with external inputs before [6, 31, 32]. However, there are some application-specific choices in the implementation, so for completeness we describe the full procedure here.
The EM algorithm seeks to maximize the log-posterior of the parameters given the choice and input data, . The log-posterior is given, up to an unknown constant, by:
(5) |
where s indexes the session in which the data was collected (out of S total sessions), and the sum over zs is over all possible latent state paths in the Ts trials of session s.
The prior distribution over the model parameters p(Θ) that we used was:
(6) |
The prior over the GLM weight vectors wk was thus an independent, zero-mean Gaussian with variance σ2. Smaller values of σ2 have the effect of shrinking the fitted weights towards zero, while σ2 = ∞ corresponds to a flat prior. For the transition matrix, we placed an independent Dirichlet prior over each row Ai, which is a natural choice for vectors on the unit (K − 1) simplex (i.e., all elements of the vector must be non-negative and sum to 1). The Dirichlet distribution is controlled by a shape parameter . A value of α = 1 corresponds to a flat prior, while large values of α have the effect of spreading probability mass equally over states. We also placed a Dirichlet prior over the initial state distribution π.
In order to select the hyperparameters σ and α governing the prior, we performed a grid search for σ ∈ {0.5, 0.75, 1, 2, 3} and α ∈ {1, 2} and selected the hyperparameters that resulted in the best performance on a held-out validation set. For IBL mice, the prior hyperparameters selected were σ = 2 and α = 2. For mice in the Odoemene et al. [20] dataset, the best hyperparameters were σ = 0.75 and α = 2. Finally, for the initial state distribution, we set απ = 1.
4.1.2. Expectation Maximization Algorithm
We used the Expectation Maximization (EM) algorithm [61, 62] to maximize the log posterior given in Eq. 5 with respect to the GLM-HMM parameters. As the sum given in Eq. 5 involves an exponential number of terms— terms to be specific—we don’t maximize this expression directly. Instead, the EM algorithm provides an efficient way to compute this term using a single forward and backward pass over the data. During the E-step of the EM algorithm we compute the “Expected Complete-Data log-likelihood” (ECLL), which is a lower bound on the right-hand side of Eq. 5 [61–63]. Then, during the ‘Maximization’ or M-step of the algorithm, we maximize the ECLL with respect to the model parameters, Θ. It can be shown that this procedure has the effect of always improving the log posterior in each step of the algorithm [63].
Concretely, the ECLL can be written as a sum over the ECLLs for each session and the log-prior. The ECLL for a single session s can be written
(7) |
In order to get to the second line, we substituted the definition of the joint distribution for the GLM-HMM:
(8) |
In Eq. 7, p(ys,t|zs,t = k, xs,t, wk) is the Bernoulli GLM distribution given by Eq. 4. Finally,
(9) |
is the posterior state probability at trial t (in session s) for state k, while
(10) |
is the joint posterior state distribution for two consecutive latents.
While the formula for the log posterior (Eq. 5) involved the sum over all possible state assignments at each trial, by taking advantage of the structure of the joint probability distribution for the GLM-HMM (Eq. 8), this sum was implemented efficiently and Eq. 7 involves summing over at most O(K2Ts) elements.
The single and joint posterior state probabilities, γs,t,k and ξs,t,k, are estimated via the forward-backward algorithm, which we describe below.
4.1.3. Expectation Step: Forward-Backward Algorithm
During the E-Step of the EM algorithm, the single and joint posterior state probabilities for all trials and states, {γs,t,k} and {ξs,t,j,k}, are estimated using the forward-backward algorithm [64] at the current setting of the GLM-HMM parameters, Θold.
The forward-backward algorithm makes use of recursion and memoization to allow these posterior probabilities to be calculated efficiently, with the forward and backward passes of the algorithm each requiring just a single pass through all trials within a session.
The goal of the forward pass is to obtain, for each trial t within session s and each state k, the quantity
(11) |
which represents the posterior probability of the choice data up until trial t and the latent state at trial t being state k.
The posterior probability associated with trial 1, as,1,k, can be calculated as follows:
(12) |
where p(ys,1|zs,1 = k, xs,1, wk) is the usual Bernoulli GLM distribution.
For trials 1 < t ≤ Ts, we can obtain these probabilities as follows:
(13) |
During the backward pass, the goal is to calculate the posterior probability of the choice data beyond the current trial, bs,t,k, for each trial t within session s and for all states, k:
(14) |
Similar to the forward pass, these quantities can be calculated efficiently by recognizing that
(15) |
and, for t ∈ {Ts − 1, …, 1}:
(16) |
From the as,t,k and bs,t,k quantities obtained via the forward-backward algorithm, we can form the single and joint posterior state probabilities, γs,t,k and ξs,t,j,k, that we actually care about for forming the ECLL of Eq. 7 as follows:
(17) |
Similarly,
(18) |
where, once again, p(ys,t+1|zt+1 = k, xt+1, wk) is the Bernoulli GLM distribution.
4.1.4. Maximization Step
After running the forward-backward algorithm, we can compute the total ECLL by summing over the per-session ECLLs (Eq. 7) and adding the log prior. During the M-step, we maximize the ECLL with respect to the GLM-HMM parameters, Θ. For the initial state distribution π and the transition matrix A, this results in the closed-form updates:
(19) |
(20) |
For the GLM weights, there is no such closed form update, but a Bernoulli GLM falls into the class of functions mapping external inputs to HMM emission probabilities considered in [31], so we know that the ECLL is concave in the GLM weights. As such, we can numerically find the GLM weights that maximize (not just locally but globally) the ECLL using the BFGS algorithm [65–68] as implemented by the scipy optimize function in python [69].
4.1.5. Comparing states across animals and GLM-HMM parameter initialization
In Fig. 4 and Fig. 5, we show the results from fitting a single GLM-HMM to each animal; however it is nontrivial to map the retrieved states across animals to one another. As such, we employed a multistage fitting procedure that allowed us to make this comparison, and we detail this procedure in Algorithm 1. In the first stage, we concatenated the data for all animals in a single dataset together (for the IBL dataset, this would be the data for all 37 animals). We then fit a GLM (a 1 state GLM-HMM) to the concatenated data using Maximum Likelihood estimation. We used the fit GLM weights to initialize the GLM weights of a K-state GLM-HMM that we again fit to the concatenated dataset from all animals together (to obtain a “global fit”). We added Gaussian noise with σinit = 0.2 to the GLM weights, so that the initialized states were distinct, and we initialized the transition matrix of the K-State GLM-HMM as where and . We then normalized this so that that rows of the transition matrix added up to 1, and represented probabilities. While the EM algorithm is guaranteed to converge to a local optimum in the log probability landscape of Eq. 5, there is no guarantee that it will converge to the global optimum [70]. Correspondingly, for each value of K, we fit the model 20 times using 20 different initializations.
In the next stage of the fitting procedure, we wanted to obtain a separate GLM-HMM fit for each animal, so we initialized a model for each animal with the GLM-HMM global fit parameters from all animals together (out of the 20 initializations, we chose the model that resulted in the best training set log-likelihood). We then ran the EM algorithm to convergence; it is these recovered parameters that are shown in Fig. 4 and Fig. 5. By initializing each individual animal’s model with the parameters from the fit to all animals together, it was no longer necessary for us to permute the retrieved states from each animal so as to map semantically similar states to one another.
We note that the initialization scheme detailed above is sufficiently robust so as to allow recovery of GLM-HMM parameters in various parameter regimes of interest. In particular, we simulated datasets from a GLM-HMM with the global fit parameters for both the IBL and Odoemene et al. [20] datasets, as well as a global fit lapse model. We show the results of these recovery analyses in Fig. ED9 and Fig. ED10.
4.2. Assessing Model Performance
4.2.1. Cross Validation
There are two ways in which to perform cross-validation when working with Hidden Markov Models. Firstly, it is possible to hold out entire sessions of choice data for assessing test-set performance. That is, when fitting the model, the objective function in Eq. 5 and the ECLL in Eq. 7 are modified to only include 80% of sessions (since we use 5-fold cross-validation throughout this work); and the log-likelihood of the held-out 20% of sessions is calculated using the fit parameters, and a single run of the forward-pass on the held-out sessions:
(21) |
where S \ S′ is the set of held out sessions, and Θ′ is the set of GLM-HMM parameters obtained by fitting the model using the trials from S′.
The second method of performing cross-validation involves holding out 20% of trials within a session. When fitting the model, the third term in the ECLL is modified so as to exclude these trials and is now , where T′ is the set of trials to be used to fit the model. Furthermore, the calculation of the posterior state probabilities, γs,t,k and ξs,t,j,k, is also modified so as to exclude the test set choice data. In particular, γs,t,k is now and similarly ξs,t,j,k is now . The method of calculating these modified posterior probabilities is as detailed in Eq. 17 and Eq. 18, but now the calculation of the forward and backward probabilities, as,t,k and bs,t,k in Eq. 12, Eq. 13, Eq. 15 and Eq. 16 is modified so that, on trials that are identified as test trials, the p(ys,t|zs,t = k, xs,t, wk) term in these equations is replaced with 1.
In Fig. 2, Fig. 4 and Fig. 5, we perform cross-validation by holding out entire sessions. We believed it would be harder to make good predictions on entire held out sessions, compared to single trials within a session, as we thought that mice would exhibit more variability in behavior across sessions compared to within sessions. When we compare the performance of the GLM-HMM against the PsyTrack model of [35] in Fig. 4f, we use the second method of cross-validation so as to use the same train and test sets as PsyTrack (PsyTrack cannot make predictions on entire held-out sessions).
4.2.2. Units for test set log-likelihood
In Fig. 2, Fig. 4, Fig. 5 and Fig. 7, as well as in some supplementary figures, we report the log-likelihood of different models on held-out sessions in units of bits per trial. This is calculated as follows:
(22) |
where, for the GLM-HMM, LLtest is the log-likelihood of the entire test set as calculated in Eq. 21, and LL0 is the log-likelihood of the same test set under a Bernoulli model of animal choice behavior. Specifically, this baseline model assumes that animals flip a coin on each trial so as to decide to go Right, and the probability of going Right is equal to the fraction of trials in the training set in which the animal chose to go to the Right. ntest is the number of trials in the test set, and is important to include since LLtest depends on the number of trials in the test set. Dividing by log(2) gives the log-likelihood the units of bits per trial. Clearly, larger values of the log-likelihood are better, with a value of 0 indicating that a model offers no improvement in prediction compared to the crude baseline model described above. However, even small values of test set log-likelihood, when reported in units of bits per trial, can indicate a large improvement in predictive power. For a test set size of ntest = 500, a log-likelihood value of 0.01 bits/trial indicates that the test data is 31.5 times more likely to have been generated with the GLM-HMM compared to the baseline model. For a test set of ntest = 5000, and the same value of 0.01 bits/trial for the log-likelihood, the test set becomes 1×1015 times more likely under the GLM-HMM compared to the baseline model!
4.2.3. Predictive Accuracy
In Fig. 2, Fig. 4 and Fig. 5, we also report the predictive accuracy of the GLM-HMM. When calculating the predictive accuracy, we employ a method similar to the second method described above in section 4.2.1. In particular, we hold out 20% of trials and then obtain the posterior state probabilities for these trials, t″ ∈ {T \ T′}, as, using the other 80% of trials (this latter set of trials being labeled T′). We then calculate the probability of the held-out choices being to go Right as:
(23) |
We then calculate the predictive accuracy as:
(24) |
4.3. Comparison with PsyTrack Model of Roy et al.
The PsyTrack model of Roy et al. [36] assumes an animal makes its choice at trial t according to
(25) |
where wt evolves according to
(26) |
where and . Specifically, the animal is assumed to use a set of slowly changing weights to make its decision on each trial.
In order to perform model comparison with the PsyTrack model of [35, 36], we utilized the code provided at https://github.com/nicholas-roy/psytrack (version 1.3).
4.3.1. Simulating Choice Data with AR(1) Model for Weights
While the GLM-HMM is a generative model that can be readily used to simulate choice data that resembles, in accuracy and in the resulting psychometric curves, the choice data of real animals, this is not true for the PsyTrack model of [35, 36]. Indeed, specifying only the , hyperparameters of that model and then generating weights according to Eq. 26 and choice data according to Eq. 25 will likely result in choice behavior that is vastly different from that of real animals (the PsyTrack model is underconstrained as a generative model). As such, so as to produce the rightmost panel of Fig. 4f, we simulated smoothly evolving weights from an AR(1) model where the parameters of this model were obtained using the PsyTrack fits to real data. Specifically, we assumed that the probability of going rightward at time t was given by:
(27) |
and we assumed that the weights evolved according to an AR(1) process as follows:
(28) |
where wm,t is the mth element of wt in Eq. 27 and is the average weight that an animal places on covariate m across all trials when fit with the PsyTrack model: . We obtained αm by regressing against and taking the retrieved slope (after confirming that the retrieved slope had a magnitude of less than 1, so as to ensure that the weights did not diverge as t → ∞). , where and was obtained from the PsyTrack fit to an animal’s real choice data. Finally, we set w0 = 0.
We simulated weight trajectories for thousands of trials for each animal, so that the AR(1) process reached the stationary regime for each covariate, and so that the mean, variance and autocovariance of each weight for each covariate were close to those returned by the PsyTrack fits to the real choice data.
4.4. Datasets studied
In this paper, we applied the GLM-HMM to three publicly available behavioral datasets associated with recent publications. Firstly, we studied the data associated with [19] that is made available via figshare at https://doi.org/10.6084/m9.figshare.11636748. We used the framework developed in [71] to access the data. We modeled the choice data for the 37 animals in this dataset which had more than 30 sessions of data during the ‘bias block’ regime. We focused on this regime because of the fact that mice, when they have reached this regime, understand the rules of the task and exhibit stationary behavior (see Fig. S1 for plots of accuracy against session identity for each animal, as well as Fig. S2 for the psychometric curves for these animals for the trials studied). For each session, we subset to the first 90 trials of data because, during these trials, the stimulus was equally likely to appear on the left or right of the screen. After the first 90 trials, the structure of the task changed and for a block of trials, the stimulus appeared on the left with a probability of either 80% or 20%; the block identity switched multiple times throughout a session, so that 80% and 20% blocks were interleaved. We subset to the animals with more than 30 sessions of data because we were able to confidently recover GLM-HMM and lapse model generative parameters when we simulated datasets with this number of trials: see Figures Fig. ED9 and Fig. ED10. As a sanity check to make sure that the recovered states and transitions were not a consequence of the animals we study having been exposed to bias blocks in earlier sessions, we obtained data for 4 animals that were never exposed to bias blocks (not included in the publicly released dataset) and fit the GLM-HMM to the choice data for these animals. In Fig. ED3, we show that the retrieved states, dwell times and model comparison results for these animals look very similar to those shown in Fig. 4.
The second dataset that we studied was that associated with [20], with the data being made available at https://doi.org/10.14224/1.38944. Once again, we studied sessions after animals had learned the task (see Fig. S13 and Fig. S14). For this dataset, the retrieved states were less distinct compared to those for the IBL dataset, and as such, we required more trials to be able to recover the generative parameters in simulated data: see Fig. ED9. We thus subset to the 15 animals with more than 20 sessions of data and 12,000 trials of data. Compared to the IBL dataset, where the violation rate across all animals’ data was less than 1% of trials (where a violation is where the animal chose not to respond), the violation rate across the 15 animals that we studied from this second dataset was 21%. Thus, it was important to develop a principled method for dealing with violation trials. We treated violation trials as trials with missing choice data, and we handled these trials in a similar way to how we handled test data when performing the second type of cross-validation described in section 4.2.1 above. That is, we modified the third term of the ECLL given in Eq. 7 to exclude violation trials, and we modified the definition of the posterior state probabilities for these trials to be and , where T′, rather than representing the training set data, is now the set of non-violation trials. The calculation of the forward and backward probabilities, as,t,k and bs,t,k, was modified so that, on violation trials, in Eq. 12, Eq. 13, Eq. 15 and Eq. 16, the p(ys,t|zs,t = k, xs,t, wk) term was replaced with 1.
Finally, the third dataset that we studied was the human dataset associated with [27], with the data being made available at https://doi.org/10.6084/m9.figshare.4300043. We applied the GLM-HMM to the data from all 27 participants in this dataset. This dataset has 5 sessions, each of approximately 500 trials, for each participant and does not include the initial session that was used to teach participants how to perform this task.
4.4.1. Forming the Design Matrix
Each of the models discussed in this paper (GLM-HMM, the classic lapse model, the PsyTrack model of Roy et al. [36]) were fit using a design matrix of covariates, , where T was the number of trials of choice data for a particular animal. A single row in this matrix was the vector of covariates, , influencing the animal’s choice at trial t. For all analyses presented in text, unless specified otherwise, M = 4.
For both tasks, the first column in the design matrix was the z-scored stimulus intensity. For the IBL task, we calculated the stimulus intensity as the difference in the value of the visual contrast on the right side of the screen minus the visual contrast on the left of the screen. This resulted in 9 different values for the ‘signed contrast’: {−100, −25, −12.5, −6.25, 0, 6.25, 12.5, 25, 100}. We then z-scored this difference quantity across all trials. For the Odoemene et al. [20] task, we subtracted the 12Hz threshold from the flash rate presented on each trial and then z-scored the resulting quantity. For the human dataset [27], we z-scored the relative motionstrength presented on each trial.
For all trials, all animals and both tasks, the second column of the design matrix was set to 1, so as to enable us to capture the animal’s innate bias for going rightward or leftward. The third column in the design matrix was the animal’s choice on the previous trial . Whereas yt−1 ∈ {0, 1}, Xt,3 ∈ {−1, 1}. It is not strictly necessary to perform this scaling, but we did so to ensure that the range of values for X:,1 and X:,3 were more similar (which can be useful when performing parameter optimization). Finally, the fourth column in the design matrix was the win-stay-lose-switch covariate, which was calculated as , where rt−1 ∈ {−1, 1} was a binary variable indicating whether or not the animal was rewarded on the previous trial. Again X:,4 ∈ {−1, 1}.
4.4.2. Response time data
In Fig. 6, we analyzed response times for IBL animals, as reaction time data was unavailable for most sessions. The mouse’s response time is the time from the stimulus onset to the time at which it received feedback on its decision (receiving its reward/hearing an auditory cue to indicate an error trial). In comparison, a reaction time for the IBL task is the time from the stimulus appearing on the screen to the mouse moving the wheel for the first time in the direction of its decision. Response times ranged from a few hundred milliseconds to 10 seconds in the most extreme case, while typical reaction times (calculated on a few sessions for which we had reliable data) tended to be on the order of a few hundred milliseconds. Relative to other tasks, the IBL task is a task that results in short reaction times.
4.4.3. Data Availability
The raw data studied in this paper is publicly available: the IBL data associated with [19] can be accessed at https://doi.org/10.6084/m9.figshare.11636748. The Odoemene et al. data associated with [20] can be accessed at https://doi.org/10.14224/1.38944. Finally, the human data associated with [27] can be accessed at https://doi.org/10.6084/m9.figshare.4300043.
4.5. Code availability
We contributed code to version 0.0.1 of the Bayesian State Space Modeling framework of [72] and we use this code base to perform GLM-HMM inference. The code to analyze the resulting model fits and to produce the figures in this paper is available at https://github.com/zashwood/glm-hmm.
4.6. Statistics and Reproducibility
We examined the choice data of animals in the IBL dataset with more than 3000 trials of data. In the case of Odoemene et al. animals, we required 12,000 trials of data. As detailed in the Methods section, as well as in the supplementary information, we determined that these numbers were sufficient because we found that we could successfully recover the generative parameters of a GLM-HMM in simulated data in the two different parameter regimes with these numbers of trials. All animals with the requisite number of trials of data were included in the analyses we present in the paper. Each experiment presented in the paper was repeated in multiple animals (37 in the case of the IBL dataset, 15 for the Odoemene et al. dataset and 27 in the case of the Urai et al. human dataset). The effects identified were largely consistent across subjects: the choice data for all 37 IBL animals, 15 Odoemene et al. mice and for 24 out of the 27 humans were better explained by a GLM-HMM compared to a classic lapse model. Analysis was performed with code that is freely available to promote replication. We randomized the allocation of sessions to folds for each animal when performing 5-fold cross-validation. Blinding was not performed. All trials from animals with the requisite number of trials were analyzed; blinding was not necessary as cross-validation ensured that the results were not due to the particular choice of train/test split.
Extended Data
Supplementary Material
6. Acknowledgments
We are grateful to Miles Wells, Rebecca Terry, Laura Funnell and the Cortexlab at University College London for providing us with the data for the 4 mice plotted in Fig. ED3. We are grateful to Scott Linderman for developing the beautiful Bayesian State Space Modeling framework of [72]: as described in our Methods section, we built our code on top of this framework. We thank members of the Pillow Lab, the International Brain Laboratory (IBL), and specifically the Behavior Analysis Working Group within the IBL for helpful feedback throughout the project. We thank Peter Dayan, Sebastian Bruijns and Liam Paninski for acting as the IBL Review Board for this paper. We thank Emily Dennis for feedback at various points during the project. We thank Abigail Russo and Matthew Whiteway for providing feedback on drafts of this manuscript. We thank Hannah Bayer for her help and advice as we were preparing to submit this paper. Finally, we thank the anonymous reviewers for their insightful comments; our manuscript is greatly improved as a result of your input.
This work was supported by grants SCGB AWD543027 (J.W.P.) and SCGB AWD543011 (A.P.) from the Simons Collaboration on the Global Brain, grants NS104899 (J.W.P.), R01EB026946 (J.W.P.) and R01EY022979 (A.K.C.) from the NIH BRAIN initiative, U19 NIH-NINDS BRAIN Initiative Award 5U19NS104648 (J.W.P.), grant 315230_197296 from the Swiss National Fund (A.P.) and Wellcome Trust grants 209558 (the IBL) and 216324 (the IBL). A.E.U. was supported by the German National Academy of Sciences Leopoldina and the International Brain Research Organization. The funders had no role in study design, data collection and analysis, decision to publish or preparation of the manuscript.
5. The International Brain Laboratory
The following authors are members of the International Brain Laboratory: Zoe C Ashwood1, 2, Nicholas A. Roy2, Anne E. Urai3, Anne K. Churchland4, Alexandre Pouget5 and Jonathan W. Pillow2, 6.
Footnotes
Competing Interests Statement
The authors declare no competing interests.
References
- 1.Gomez-Marin A, Paton JJ, Kampff AR, Costa RM & Mainen ZF Big behavioral data: psychology, ethology and the foundations of neuroscience. en. Nature Neuroscience 17. Number:11 Publisher: Nature Publishing Group, 1455–1462. ISSN: 1546–1726. https://www.nature.com/articles/nn.3812 (2020) (Nov. 2014). [DOI] [PubMed] [Google Scholar]
- 2.Krakauer JW, Ghazanfar AA, Gomez-Marin A, MacIver MA & Poeppel D Neuroscience Needs Behavior: Correcting a Reductionist Bias. Neuron 93, 480–490. ISSN: 0896–6273. http://www.sciencedirect.com/science/article/pii/S0896627316310406 (2019) (Feb. 2017). [DOI] [PubMed] [Google Scholar]
- 3.Berman GJ Measuring behavior across scales. BMC Biology 16, 23. ISSN: 1741–7007. 10.1186/s12915-018-0494-7 (2020) (Feb. 2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Wiltschko AB et al. Mapping Sub-Second Structure in Mouse Behavior. English. Neuron 88. Publisher: Elsevier, 1121–1135. ISSN: 0896–6273. https://www.cell.com/neuron/abstract/S0896-6273(15)01037-5 (2020) (Dec. 2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Sharma A, Johnson R, Engert F & Linderman S in Advances in Neural Information Processing Systems 31 (eds Bengio S et al. ) 10919–10930 (Curran Associates, Inc., 2018). http://papers.nips.cc/paper/8289-point-process-latent-variable-models-of-larval-zebrafish-behavior.pdf (2020). [Google Scholar]
- 6.Calhoun AJ, Pillow JW & Murthy M Unsupervised identification of the internal states that shape natural behavior. en. Nature Neuroscience 22. Number: 12 Publisher: Nature Publishing Group, 2040–2049. ISSN: 1546–1726. https://www.nature.com/articles/s41593-019-0533-x (2020) (Dec. 2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Batty E et al. in Advances in Neural Information Processing Systems 32 (eds Wallach H et al. ) 15706–15717 (Curran Associates, Inc., 2019). http://papers.nips.cc/paper/9701-behavenet-nonlinear-embedding-and-bayesian-neural-decoding-of-behavioral-videos.pdf (2020). [Google Scholar]
- 8.Johnson RE et al. Probabilistic Models of Larval Zebrafish Behavior Reveal Structure on Many Scales. en. Current Biology 30, 70–82. e4. ISSN: 0960–9822. http://www.sciencedirect.com/science/article/pii/S0960982219314654 (2020) (Jan. 2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Green DM & Swets JA Signal detection theory and psychophysics (Wiley New York, 1966). [Google Scholar]
- 10.Klein SA Measuring, estimating, and understanding the psychometric function: A commentary. Perception & psychophysics 63, 1421–1455 (2001). [DOI] [PubMed] [Google Scholar]
- 11.Bogacz R, Brown E, Moehlis J, Holmes P & Cohen JD The physics of optimal decision making: a formal analysis of models of performance in two-alternative forced-choice tasks. Psychological review 113. Publisher: American Psychological Association, 700 (2006). [DOI] [PubMed] [Google Scholar]
- 12.Gold JI & Shadlen MN The neural basis of decision making. Annual review of neuroscience 30 (2007). [DOI] [PubMed] [Google Scholar]
- 13.Ratcliff R & McKoon G The diffusion decision model: theory and data for two-choice decision tasks. Neural computation 20. Publisher: MIT Press, 873–922 (2008). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Iigaya K, Fonseca MS, Murakami M, Mainen ZF & Dayan P An effect of serotonergic stimulation on learning rates for rewards apparent after long intertrial intervals. en. Nature Communications 9. Number: 1 Publisher: Nature Publishing Group, 2477. ISSN: 2041–1723. (2021) (June 2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Urai AE, de Gee JW, Tsetsos K & Donner TH Choice history biases subsequent evidence accumulation. eLife 8 (eds Verstynen T, Shinn-Cunningham BG & Verstynen T) Publisher: eLife Sciences Publications, Ltd, e46331. ISSN: 2050–084X. 10.7554/eLife.46331 (2021) (July 2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Lak A et al. Reinforcement biases subsequent perceptual decisions when confidence is low, a widespread behavioral phenomenon. eLife 9 (eds Salinas E, Frank MJ, Salinas E, Brody CD & Ding L) Publisher: eLife Sciences Publications, Ltd, e49834. ISSN: 2050–084X. 10.7554/eLife.49834 (2021) (Apr. 2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Pisupati S, Chartarifsky-Lynn L, Khanal A & Churchland AK Lapses in perceptual decisions reflect exploration. eLife 10, e55490 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Beron C, Neufeld S, Linderman S & Sabatini B Efficient and stochastic mouse action switching during probabilistic decision making. bioRxiv. Publisher: Cold Spring Harbor Laboratory, preprint: https://www.biorxiv.org/content/early/2021/05/14/2021.05.13.444094.full.pdf. https://www.biorxiv.org/content/early/2021/05/14/2021.05.13.444094 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.International Brain Laboratory, T. et al. A standardized and reproducible method to measure decision-making in mice. en. bioRxiv, 2020.01.17.909838. https://www.biorxiv.org/content/10.1101/2020.01.17.909838v2 (2020) (Jan. 2020). [Google Scholar]
- 20.Odoemene O, Pisupati S, Nguyen H & Churchland AK Visual Evidence Accumulation Guides Decision-Making in Unrestrained Mice. en. Journal of Neuroscience 38, 10143–10155. ISSN: 0270–6474, 1529–2401. https://www.jneurosci.org/content/38/47/10143 (2019) (Nov. 2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Pinto L et al. An Accumulation-of-Evidence Task Using Visual Pulses for Mice Navigating in Virtual Reality. English. Frontiers in Behavioral Neuroscience 12. Publisher: Frontiers. ISSN: 1662–5153. https://www.frontiersin.org/articles/10.3389/fnbeh.2018.00036/full?report=reader (2020) (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Wichmann FA & Hill NJ The psychometric function: I. Fitting, sampling, and goodness of fit. en. Perception & Psychophysics 63, 1293–1313. ISSN: 1532–5962. 10.3758/BF03194544 (2020) (Nov. 2001). [DOI] [PubMed] [Google Scholar]
- 23.Prins N The psychometric function: The lapse rate revisited. en. Journal of Vision 12, 25–25. ISSN: 1534–7362. https://jov.arvojournals.org/article.aspx?articleid=2192299 (2019) (June 2012). [DOI] [PubMed] [Google Scholar]
- 24.Carandini M & Churchland AK Probing perceptual decisions in rodents. eng. Nature Neuroscience 16, 824–831. ISSN: 1546–1726 (July 2013). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Lindsay BG Mixture Models: Theory, Geometry and Applications. NSF-CBMS Regional Conference Series in Probability and Statistics 5. Publisher: Institute of Mathematical Statistics, i–163. ISSN: 1935–5920. https://www.jstor.org/stable/4153184 (2020) (1995). [Google Scholar]
- 26.Law C-T & Gold JI Reinforcement learning can account for associative and perceptual learning on a visual-decision task. en. Nature Neuroscience 12. Number: 5 Publisher: Nature Publishing Group, 655–663. ISSN: 1546–1726. https://www.nature.com/articles/nn.2304 (2020) (May 2009). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Urai AE, Braun A & Donner TH Pupil-linked arousal is driven by decision uncertainty and alters serial choice bias. Nature Communications 8. Bandiera_abtest: a Cc_license_type: cc_by Cg_type: Nature Research Journals Number: 1 Primary_atype: Research Publisher: Nature Publishing Group Subject_term: Decision;Human behaviour Subject_term_id: decision;human-behaviour, 14637. ISSN: 2041–1723. https://www.nature.com/articles/ncomms14637 (2021) (Mar. 2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Busse L et al. The Detection of Visual Contrast in the Behaving Mouse. en. Journal of Neuroscience 31, 11351–11361. ISSN: 0270–6474, 1529–2401. http://www.jneurosci.org/content/31/31/11351 (2019) (Aug. 2011). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Fründ I, Wichmann FA & Macke JH Quantifying the effect of intertrial dependence on perceptual decisions. en. Journal of Vision 14. Publisher: The Association for Research in Vision and Ophthalmology, 9–9. ISSN: 1534–7362. https://jov.arvojournals.org/article.aspx?articleid=2194025 (2020) (June 2014). [DOI] [PubMed] [Google Scholar]
- 30.Lueckmann J-M, Macke JH & Nienborg H Can Serial Dependencies in Choices and Neural Activity Explain Choice Probabilities? en. Journal of Neuroscience 38. Publisher: Society for Neuroscience Section: Research Articles, 3495–3506. ISSN: 0270–6474, 1529–2401. https://www.jneurosci.org/content/38/14/3495 (2021) (Apr. 2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Escola S, Fontanini A, Katz D & Paninski L Hidden Markov models for the stimulus-response relationships of multistate neural systems. Neural computation 23, 1071–1132 (2011). [DOI] [PubMed] [Google Scholar]
- 32.Bengio Y & Frasconi P in Advances in Neural Information Processing Systems 7 (eds Tesauro G, Touretzky DS & Leen TK) 427–434 (MIT Press, 1995). http://papers.nips.cc/paper/964-an-input-output-hmm-architecture.pdf (2020). [Google Scholar]
- 33.Burgess CP et al. High-Yield Methods for Accurate Two-Alternative Visual Psychophysics in Head-Fixed Mice. eng. Cell Reports 20, 2513–2524. ISSN: 2211–1247 (Sept. 2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Friedman J, Hastie T & Tibshirani R The elements of statistical learning 10 (Springer series in statistics New York, 2001). [Google Scholar]
- 35.Roy NA, Bak JH, Akrami A, Brody C & Pillow JW Efficient inference for time-varying behavior during learning in Advances in Neural Information Processing Systems (2018), 5695–5705. [PMC free article] [PubMed]
- 36.Roy NA et al. Extracting the dynamics of behavior in sensory decision-making experiments. eng. Neuron 109, 597–610. e6. ISSN: 1097–4199 (Feb. 2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Heitz RP The speed-accuracy tradeoff: history, physiology, methodology, and behavior. English. Frontiers in Neuroscience 8. Publisher: Frontiers. ISSN: 1662–453X. https://www.frontiersin.org/articles/10.3389/fnins.2014.00150/full (2020) (2014). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Ratcliff R, Smith PL, Brown SD & McKoon G Diffusion Decision Model: Current Issues and History. Trends in cognitive sciences 20, 260–281. ISSN: 1364–6613. https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4928591/ (2020) (Apr. 2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Shevinsky CA & Reinagel P The Interaction Between Elapsed Time and Decision Accuracy Differs Between Humans and Rats. English. Frontiers in Neuroscience 13. Publisher: Frontiers. ISSN: 1662–453X. https://www.frontiersin.org/articles/10.3389/fnins.2019.01211/full#B20 (2020) (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Ding L & Gold JI Neural correlates of perceptual decision making before, during, and after decision commitment in monkey frontal eye field. eng. Cerebral Cortex (New York, N.Y.: 1991) 22, 1052–1067. ISSN: 1460–2199 (May 2012). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Ding L & Gold JI The basal ganglia’s contributions to perceptual decision making. eng. Neuron 79, 640–649. ISSN: 1097–4199 (Aug. 2013). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Engel TA et al. Selective modulation of cortical state during spatial attention. en. Science 354. Publisher: American Association for the Advancement of Science Section: Report, 1140–1144. ISSN: 0036–8075, 1095–9203. https://science.sciencemag.org/content/354/6316/1140 (2020) (Dec. 2016). [DOI] [PubMed] [Google Scholar]
- 43.Zhou Y & Freedman DJ Posterior parietal cortex plays a causal role in perceptual and categorical decisions. en. Science 365. Publisher: American Association for the Advancement of Science Section: Report, 180–185. ISSN: 0036–8075, 1095–9203. https://science.sciencemag.org/content/365/6449/180 (2021) (July 2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Steinmetz NA, Zatka-Haas P, Carandini M & Harris KD Distributed coding of choice, action and engagement across the mouse brain. en. Nature 576. Number: 7786 Publisher: Nature Publishing Group, 266–273. ISSN: 1476–4687. https://www.nature.com/articles/s41586-019-1787-x (2020) (Dec. 2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Zoltowski D, Pillow J & Linderman S A general recurrent state space framework for modeling neural dynamics during decision-making en. in International Conference on Machine Learning ISSN: 2640–3498 (PMLR, Nov. 2020), 11680–11691. https://proceedings.mlr.press/v119/zoltowski20a.html (2021). [Google Scholar]
- 46.Hennig JA et al. Learning is shaped by abrupt changes in neural engagement. en. Nature Neuroscience. Publisher: Nature Publishing Group, 1–10. ISSN: 1546–1726. https://www.nature.com/articles/s41593-021-00822-8 (2021) (Mar. 2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47.Bolkan SS et al. Strong and opponent contributions of dorsomedial striatal pathways to behavior depends on cognitive demands and task strategy. bioRxiv. Publisher: Cold Spring Harbor Laboratory; preprint: https://www.biorxiv.org/content/early/2021/07/25/2021.07.23.453573.full.pdf. https://www.biorxiv.org/content/early/2021/07/25/2021.07.23.453573 (2021). [Google Scholar]
- 48.Carter ME et al. Tuning arousal with optogenetic modulation of locus coeruleus neurons. en. Nature Neuroscience 13. Number: 12 Publisher: Nature Publishing Group, 1526–1533. ISSN: 15461726. https://www.nature.com/articles/nn.2682 (2021) (Dec. 2010). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49.Harris KD & Thiele A Cortical state and attention. en. Nature Reviews Neuroscience 12. Number: 9 Publisher: Nature Publishing Group, 509–523. ISSN: 1471–0048. https://www.nature.com/articles/nrn3084 (2021) (Sept. 2011). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.Reimer J et al. Pupil fluctuations track fast switching of cortical states during quiet wakefulness. eng. Neuron 84, 355–362. ISSN: 1097–4199 (Oct. 2014). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51.Vinck M, Batista-Brito R, Knoblich U & Cardin JA Arousal and locomotion make distinct contributions to cortical activity patterns and visual encoding. eng. Neuron 86, 740–754. ISSN: 1097–4199 (May 2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 52.Ashwood Z, Roy NA, Bak JH & Pillow JW Inferring learning rules from animal decisionmaking. en. Advances in Neural Information Processing Systems 33, 3442–3453. https://proceedings.neurips.cc/paper/2020/hash/234b941e88b755b7a72a1c1dd5022f30-Abstract.html (2021) (2020). [PMC free article] [PubMed] [Google Scholar]
- 53.Daw ND, O’Doherty JP, Dayan P, Seymour B & Dolan RJ Cortical substrates for exploratory decisions in humans. Nature 441, 876–879. ISSN: 0028–0836. https://www.ncbi.nlm.nih.gov/pmc/articles/PMC2635947/ (2020) (June 2006). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 54.Dayan P & Daw ND Decision theory, reinforcement learning, and the brain. Cognitive, Affective, & Behavioral Neuroscience 8. Publisher: Springer, 429–453 (2008). [DOI] [PubMed] [Google Scholar]
- 55.Behrens TEJ, Woolrich MW, Walton ME & Rushworth MFS Learning the value of information in an uncertain world. en. Nature Neuroscience 10. Number: 9 Publisher: Nature Publishing Group, 1214–1221. ISSN: 1546–1726. https://www.nature.com/articles/nn1954 (2021) (Sept. 2007). [DOI] [PubMed] [Google Scholar]
- 56.Fiser J, Berkes P, Orbán G & Lengyel M Statistically optimal perception and learning: from behavior to neural representations. en. Trends in Cognitive Sciences 14, 119–130. ISSN: 1364–6613. https://www.sciencedirect.com/science/article/pii/S1364661310000045 (2021) (Mar. 2010). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 57.Mendonça AG et al. The impact of learning on perceptual decisions and its implication for speed-accuracy tradeoffs. en. Nature Communications 11. Number: 1 Publisher: Nature Publishing Group, 2757. ISSN: 2041–1723. https://www.nature.com/articles/s41467-020-16196-7 (2020) (June 2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 58.Gershman SJ Origin of perseveration in the trade-off between reward and complexity. en. bioRxiv. Publisher: Cold Spring Harbor Laboratory Section: New Results, 2020.01.16.903476. https://www.biorxiv.org/content/10.1101/2020.01.16.903476v2 (2021) (Apr. 2020). [DOI] [PubMed] [Google Scholar]
- 59.Weilnhammer V, Chikermane M & Sterzer P Bistable perception alternates between internal and external modes of sensory processing. en. iScience 24, 102234. ISSN: 2589–0042. https://www.sciencedirect.com/science/article/pii/S2589004221002029 (2021) (Mar. 2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 60.Weilnhammer VA, Stuke H, Eckert A-L, Standvoss K & Sterzer P Humans and mice fluctuate between external and internal modes of sensory processing en. preprint (Neuroscience, Aug. 2021). http://biorxiv.org/lookup/doi/10.1101/2021.08.20.457079 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 61.Dempster AP, Laird NM & Rubin DB Maximum likelihood from incomplete data via the EM algorithm. Journal of the Royal Statistical Society: Series B (Methodological) 39, 1–22 (1977). [Google Scholar]
- 62.McLachlan GJ & Krishnan T The EM Algorithm and Extensions en. Google-Books-ID: NBawza-WoWa8C. ISBN: 978-0-470-19160-6 (John Wiley & Sons, Nov. 2007). [Google Scholar]
- 63.Bishop CM Pattern recognition and machine learning (springer, 2006). [Google Scholar]
- 64.Baum LE, Petrie T, Soules G & Weiss N A Maximization Technique Occurring in the Statistical Analysis of Probabilistic Functions of Markov Chains. EN. Annals of Mathematical Statistics 41, 164–171. ISSN: 0003–4851, 2168–8990. https://projecteuclid.org/euclid.aoms/1177697196 (2020) (Feb. 1970). [Google Scholar]
- 65.Broyden CG The Convergence of a Class of Double-rank Minimization Algorithms 1. General Considerations. en. IMA Journal of Applied Mathematics 6. Publisher: Oxford Academic, 76–90. ISSN: 0272–4960. https://academic.oup.com/imamat/article/6/1/76/746016 (2020) (Mar. 1970). [Google Scholar]
- 66.Fletcher R A new approach to variable metric algorithms. en. The Computer Journal 13. Publisher: Oxford Academic, 317–322. ISSN: 0010–4620. https://academic.oup.com/comjnl/article/13/3/317/345520 (2020) (Jan. 1970). [Google Scholar]
- 67.Goldfarb D A family of variable-metric methods derived by variational means. en. Mathematics of Computation 24, 23–26. ISSN: 0025–5718, 1088–6842. https://www.ams.org/mcom/1970-24-109/S0025-5718-1970-0258249-6/ (2020) (1970). [Google Scholar]
- 68.Shanno DF Conditioning of quasi-Newton methods for function minimization. en. Mathematics of Computation 24, 647–656. ISSN: 0025–5718, 1088–6842. https://www.ams.org/mcom/1970-24-111/S0025-5718-1970-0274029-X/ (2020) (1970). [Google Scholar]
- 69.Virtanen P et al. SciPy 1.0: Fundamental Algorithms for Scientific Computing in Python. Nature Methods 17, 261–272 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 70.Salakhutdinov R, Roweis ST & Ghahramani Z Optimization with EM and expectation-conjugate-gradient in Proceedings of the 20th International Conference on Machine Learning (ICML-03) (2003), 672–679. [Google Scholar]
- 71.International Brain Laboratory, T. et al. Data architecture for a large-scale neuroscience collaboration. bioRxiv. https://www.biorxiv.org/content/early/2020/02/23/827873 (2020). [Google Scholar]
- 72.Linderman S, Antin B, Zoltowski D & Glaser J SSM: Bayesian Learning and Inference for State Space Models version 0.0.1. Oct. 2020. https://github.com/lindermanlab/ssm.
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
The raw data studied in this paper is publicly available: the IBL data associated with [19] can be accessed at https://doi.org/10.6084/m9.figshare.11636748. The Odoemene et al. data associated with [20] can be accessed at https://doi.org/10.14224/1.38944. Finally, the human data associated with [27] can be accessed at https://doi.org/10.6084/m9.figshare.4300043.