Skip to main content
PLOS Biology logoLink to PLOS Biology
. 2022 Mar 31;20(3):e3001566. doi: 10.1371/journal.pbio.3001566

The effect of apathy and compulsivity on planning and stopping in sequential decision-making

Jacqueline Scholl 1,2,3,4,*, Hailey A Trier 3, Matthew F S Rushworth 3,5,‡,#, Nils Kolling 4,6,‡,#
Editor: Christopher D Chambers7
PMCID: PMC8970514  PMID: 35358177

Abstract

Real-life decision-making often comprises sequences of successive decisions about whether to take opportunities as they are encountered or keep searching for better ones instead. We investigated individual differences related to such sequential decision-making and link them especially to apathy and compulsivity in a large online sample (discovery sample: n = 449 and confirmation sample: n = 756). Our cognitive model revealed distinct changes in the way participants evaluated their environments and planned their own future behaviour. Apathy was linked to decision inertia, i.e., automatically persisting with a sequence of searches for longer than appropriate given the value of searching. Thus, despite being less motivated, they did not avoid the effort associated with longer searches. In contrast, compulsivity was linked to self-reported insensitivity to the cost of continuing with a sequence of searches. The objective measures of behavioural cost insensitivity were clearly linked to compulsivity only in the discovery sample. While the confirmation sample showed a similar effect, it did not reach significance. Nevertheless, in both samples, participants reported awareness of such bias (experienced as “overchasing”). In addition, this awareness made them report preemptively avoiding situations related to the bias. However, we found no evidence of them actually preempting more in the task, which might mean a misalignment of their metacognitive beliefs or that our behavioural measures were incomplete. In summary, individual variation in distinct, fundamental aspects of sequential decision-making can be linked to variation in 2 measures of behavioural traits associated with psychological illness in the normal population.


This preregistered study investigated individual differences in a sequential decision making task. The cognitive model revealed distinct differences in the way participants evaluated their environments and planned their own future behaviour connected to clinical dimensions; apathy was linked to decision inertia and compulsivity to perceived insensitivity to the cost of continuing with a sequence of searches, but also attempts of pre-emptive avoidance.

Introduction

Real-life decision-making behaviour often involves planning a sequence of decisions and then carrying them out while monitoring their appropriateness. Such sequential decision-making patterns, rather than single event decisions, often characterise the decision-making problems that have to be solved by foraging animals and, equally, by humans. While sequential decisions are ubiquitous in nature, very little attention has been paid to cognitive models of decision-making that take sequential context and planning into account. This is an important issue as many decision problems, and the cognitive mechanisms needed to deal with them, are unique to sequential decisions [13]. We illustrate an example of a sequential job search in Fig 1. Psychologically and neurally, the processes involved here are distinct from more commonly studied single event decisions [410]. It has been argued that this kind of paradigm is more ecological and captures aspects of decisions in real life particularly well [1113]. If this is the case, then it is possible that such tasks will probe individual differences in decision-making that are related to meaningful individual differences in psychological traits that, in the extreme, are related to psychological illness.

Fig 1. Sequential search paradigm.

Fig 1

(Ai) Deciding whether to leave a job can be seen as a sequential search problem in which one repeatedly encounters alternative job offers, one at a time, that one can then either accept or instead reject in order to search for another alternative from the job market. In short, the decision-maker needs to weigh up the value of the current offer with the value of the job market. The job market’s value has 2 important components. First, it consists of the average (“myopic value”) value of the alternatives in the job market [14]. Second, it consists of its “prospective value,” i.e., the value that stems from the possibility to make sequences of searches within this pool of alternatives [14]. Taking prospective value into account means that the value of the alternatives is higher than might initial be thought on the basis of the average, myopic value. The degree to which this is the case depends on how many opportunities one has to explore the alternatives and the cost of doing this (see S1 Text “Decision tree model to derive prospective value” for more details). In the illustrated sequence, searches were continued until a sufficiently good alternative was encountered (one has found one’s “dream job”). Instead, if the opportunities for exploring are limited (i.e., only a limited number of searches are available), searches might have to stop when the available opportunities to search are depleted, before a “dream job” was found. (B) Sequential decisions can be broken down into 2 stages. The first is the initial consideration of the environment and its potential outcomes. Second, the agent goes through the actual sequence of searches. Here, biases of insensitivity to search cost or decision inertia can emerge. If agents are aware of these biases, they can be preemptively avoided by not engaging with sequences in the first place especially if the potential sequence is of the type in which one may be more prone to the biases or in which the biases may be more costly. (C) Illustration of the actual task screen participants saw. The wheel of fortune shows all possible alternatives and the sizes of the areas represent their probabilities of being drawn. Participants can accept the initial offer or any later offer by clicking “bank win” or search for an alternative by deciding to “spin again” that will reduce the number of “remaining spins” by one and cost as many points as show as “cost to spin.” (D) Prospective value is derived from a decision tree model (see S1 Text). This model captures the way in which prospective value is higher in environments with many available searches (num. searches), higher variability in how good and likely the offers are (STD), particularly if many searches are available (STD*Num. searches) and low cost, again particularly if many searches are available. Results are regression weights relating prospective value to different features of the environment (see Methods). The proportion of variance of prospective value explained by the regressors (R2) was 79%. (E) Participants’ behaviour (confirmation sample—same results were found in the discovery sample) and model predictions for decisions after the initial planning stage: The higher the relative value favouring searching versus accepting the offer, the more likely participants were to search (dots with error bars). The model’s behaviour matches participants’ very closely (line). (F) Illustration of the main behavioural measures in the confirmation sample used to relate clinical factors to. In general, the more often participants (dots with error bars; continuous lines show model predictions) have already searched on a given trial, the more likely they were to search again, rather than accept an offer (“decision inertia”). Note that in contrast to the regression models (S1 and S2 Figs), this split does not correct for the fact that other factors such as search value are partially correlated with the number of previous searches, but it does provide a simple way to visualise the impact of task features on participant behaviour. The graph also shows the higher the cost (grey to blue), the less likely participants were to search. Error bars show the standard error of the mean, size of the dots show the mean number of data points per participant, and lines show the model predictions. Data in panel E are available in S1 Data and panel F in S8 Data. STD, standard deviation.

To examine this possibility, here, we have paired a previously established “sequential search task” [14] with measures of apathy and compulsivity—traits that are common features of psychological illness collected in a large online sample in a transdiagnostic approach [15]. In addition to the behavioural measures intrinsic to the sequential task, we have also collected a set of introspective and self-report measures relating to the task. Crucially, we do not just examine raw behavioural scores, but, instead, we use a computational model to obtain precise and quantitative estimates of the cognitive processes that led to the participants’ behaviour in the task. They capture firstly how well participants plan a sequence of searches and secondly how they then carry out the sequence of searches. We can obtain measures of both rational and irrational behaviour. We can measure to what extent participants can flexibly adjust their preplanned behaviours “on the fly.” We can also measure how participants fail to adjust through 2 biases (Fig 1F): First, participants can be insensitive to repeatedly incurred costs of decisions as they progress through a sequence. Second, participants can have decision inertia, i.e., inflexibility and repetitiveness. This is distinct from action inertia, which would lead to increased reaction times (RTs) but no change in which decisions are taken. Fig 1F illustrates both of these biases. Sensitivity to costs is the degree to which searching is made less likely by increasing costs. Decision inertia is the likelihood of deciding to search yet again because you have searched already; when there is decision inertia, then the higher the previous numbers of searches, then the more likely you should be to search yet again. Formally, we can use our computational models to get quantitative measures of each of these biases, while controlling for all other factors that might affect choices. Finally, we can also measure how participants’ knowledge of their own biases enables them to preemptively avoid situations that are more likely to trigger a bias. We measure this here by asking participants to self-report whether they preemptively avoided these situations.

Making categorical comparisons between patient groups and controls is an important line of research that has delivered major insights into changes in cognition and behaviour in psychological illness. An alternative approach is to use a transdiagnostic dimensionality-based approach. We chose this approach here due to its ability to extract multiple distinct dimensions within large population data and across different questionnaires and traditional diagnostic boundaries [16,17]. Specifically, we collected questionnaires with measures covering a wide range of psychiatric symptoms and then extracted symptom dimensions cutting across traditional disorder boundaries. This is particularly relevant for 3 reasons. First, many disorders, such as obsessive-compulsive disorder (OCD), have comorbidity rates as high as 90% [18]. Secondly, it is well known that some dimensions are shared across many disorders [19]. Third, recent work [20] suggests that dimensions cutting across disorders relate more strongly to behavioural deficits than diagnostic categories. Here, we are particularly interested in the dimensions of apathy and compulsivity and want to study them while controlling for other confounding factors. In general terms, the hypotheses that we want to test are as follows (for exact statistical tests, see Methods).

First, we hypothesise apathy will be associated with decision inertia. Apathy exists on a spectrum and in its more severe form is a symptom present in many disorders typically regarded as distinct. It is prominent in depression, in the negative symptoms of schizophrenia, after specific brain injuries [19], and in burnout [21]. It is sometimes described as avolition because of the problems patients can have with initiating new behaviours [22] or generating new options as candidates for decision-making [23]. If the option of stopping a sequence of decisions, such as job search, does not even occur to you then you might exhibit decision inertia, continuing to search and exerting more and more effort (i.e., “overcommitting” [24]). In this way, apathy could counterintuitively lead to the exertion of more unnecessary effort, which, in turn, might increase feelings of apathy. Therefore, we predict that apathy will be linked to an increased decision inertia (hypothesis 1), i.e., a bias to search again and again for a new offer without considering that the appropriate strategy now would be to stop the search. One could also call this a commitment or closure inertia. Note that the hypothesis is very distinct to the more simplistic idea that apathy just occurs at the action level as a failure to act quickly, which could express itself as slower response times. Instead, we are proposing it occurs at the decision level and therefore that it may even lead to the commission of more actions.

Second, we wanted to investigate effects of the compulsivity trait on sequential search behaviour. Like apathy, compulsivity exists on a spectrum and is increased in several disorders such as OCD [25], eating disorders [26], and addiction [27]. Particularly for OCD, compulsivity is seen as central: OCD is characterised by obsessions and compulsions (e.g., obsessions about cleanliness and compulsive handwashing). For an observer, such behaviour appears to lack utility, but patients with OCD carry out the compulsive behaviours despite their immense time costs [28]. One important dimension determining utility is the cost of a behaviour. Insensitivity to the costs of the compulsive behaviour could then lead to impairments in arbitration between more cognitively demanding model-based (or “goal directed”) and model-free (or “habitual”) modes of behavioural control. This leads us to our second hypothesis: OCD-related compulsivity (measured here using questions from OCD and schizotypy inventories; “compulsivity” for short) will be linked to cost insensitivity when carrying out a sequence of searches and that participants will be aware of problems linked to this bias (“overchasing,” hypothesis 2). Note that this approach is consistent with recent evidence that obsessions may not be the primary features of OCD that cause the occurrence of obsessions [2931].

Participants’ awareness of their own bias directly leads to our next question. If compulsivity is linked to an awareness of the risk of “overchasing,” do participants do anything to try to compensate for this? Interestingly, we know that OCD is an “ego-dystonic” disorder, i.e., despite awareness, a patient is not able to inhibit their obsessions or compulsions. However, sometimes, a patient may be able to avoid situations that trigger the obsession–compulsion complex altogether; for example, family members may accommodate OCD sufferers by doing the washing instead of them to avoid triggering the OCD [32]. Here, we wanted to examine the relationship between OCD symptoms and the presence of preemptive decision-making strategies that occur at the beginnings of search sequences and which prevent the person from entering situations later in the sequence in which they would be likely to perform poorly because of a bias in decision-making. We hypothesise that OCD symptoms are associated with self-reported avoidance of situations in which biases will manifest (hypothesis 3).

For data collection and analysis, we employed a “split sample” approach [33,34]. We first collected a discovery sample that we used to refine and develop hypotheses in a data-driven way. We then used a confirmation sample to test the preregistered hypotheses in a statistically rigorous way. This approach thus combines the benefits of preregistration—reduced risk of false positives due to analytical freedoms—with those of exploratory science—the analytical approach being best adapted to the specific data.

Results

A total of 830 participants completed all parts of this preregistered study (link to OSF https://osf.io/dfg2u). Exclusions: 33 participants (4%) due behavioural responses suggesting lack of task understanding; 4 participants (0.5%) due to not completing all questionnaires or failing the questionnaire attention checks and 112 participants (13%) from the behavioural analyses of initial searches; and 44 participants (5%) from later searches due to too little variation in their responses to allow analysis (see Methods). This resulted in 756 participants included for preregistered analyses linking transdiagnostic dimensions and behaviour and 795 participants included in analyses of only the questionnaires (demographics in Table 1).

Table 1. Demographics and experiment duration for participants of the confirmation sample that we included or excluded from all analyses based on questionnaire and task performance criteria (see Methods).

Excl. not enough variation in later decision responses N = 39 Excl. questionnaires N = 2 Excl. task understanding N = 33 Included N = 756 p-Value
Age 29.4 (5.99) 26.5 (7.78) 29.9 (6.14) 28.2 (6.32) 0.298
Gender 0.372
    Female 24 (61.5%) 0 (0.00%) 18 (54.5%) 385 (50.9%)
    Male 14 (35.9%) 2 (100%) 15 (45.5%) 361 (47.8%)
    Other 1 (2.56%) 0 (0.00%) 0 (0.00%) 10 (1.32%)
Education
    GCSE 2 (5.13%) 0 (0.00%) 2 (6.06%) 39 (5.16%)
    A-levels 17 (43.6%) 1 (50.0%) 7 (21.2%) 220 (29.1%)
    Bachelor 11 (28.2%) 1 (50.0%) 11 (33.3%) 319 (42.2%)
    Master’s 6 (15.4%) 0 (0.00%) 11 (33.3%) 157 (20.8%)
    Doctorate 3 (7.69%) 0 (0.00%) 2 (6.06%) 21 (2.78%)
English fluency 0.016
    Basic 1 (2.56%) 0 (0.00%) 2 (6.25%) 4 (0.53%)
    Moderate 1 (2.56%) 0 (0.00%) 5 (15.6%) 62 (8.20%)
    Native 37 (94.9%) 2 (100%) 25 (78.1%) 690 (91.3%)
Medication status 0.042
    No 37 (94.9%) 1 (50.0%) 33 (100%) 682 (90.2%)
    Yes 2 (5.13%) 1 (50.0%) 0 (0.00%) 74 (9.79%)
Experiment duration (min) 65.5 (29.5) 65.2 (9.43) 83.2 (29.0) 64.4 (18.4) <0.001
Task duration (min) 34.9 (17.8) 38.7 (1.88) 44.3 (17.5) 36.0 (11.4) 0.002
Questionnaires duration (min) 18.9 (21.3) 15.0 (2.11) 20.5 (13.9) 16.4 (7.62) 0.026
Training duration (min) 6.58 (4.50) 8.66 (5.54) 11.3 (9.07) 6.70 (6.52) 0.001

For a detailed list of medications, see S9 Table. Data shown in this table are available in S4 Data.

GCSE, General Certificate of Secondary Education.

Power analyses of the discovery sample (n = 449 for analyses linking transdiagnostic dimensions and behaviour; see link/reference for detailed results of the discovery sample data submitted for preregistration) had suggested a sample size of 750 participants to achieve 95% power for all preregistered analyses. Throughout, we report results from the confirmation sample. Analyses that deviated from the preregistration are highlighted as “exploratory.” Results of the discovery sample can be seen the preregistration (https://osf.io/dfg2u). The questionnaires from the discovery sample data have since been published as part of a meta-analysis on the relationship between apathy and compulsivity I large population samples [35].

On each of 150 trials of the task, participants chose to either “bank” an “initial offer” or carry out the decision sequence and search among the alternatives by “spinning a wheel of fortune” (Fig 1C). Their aim was to earn as many points (i.e., money) as possible. If they banked the offer (i.e., collected the number of points of the offer), they moved on to the next trial. If instead they decided to “spin,” they had to pay a small cost and could then spin the wheel to get a new offer. In each round, there was a maximum number of times participants could spin the wheel again if they were not satisfied with their draws; each time they had to pay the same cost. Participants were shown the available offers on the wheel, how likely they were (size of area on wheel), how much it costs to spin the wheel, and how often they could spin the wheel on a trial (also see Methods for more details on properties of trials). After the task, participants were asked to introspect about their performance in the task (see Methods). We split the behavioural data into initial and later searches to be able to test effects of prospection at the first decision separately from sequential decision effects such as perseverance.

Validation of sequential decision task

Detailed validation of the paradigm and description of all model parameters can be found in S1 Text and S1 and S2 Figs. The model captured participants’ behaviour well (Fig 1E and 1F). In the analyses, we related different factors that should rationally or irrationally drive behaviour (e.g., value of the offer and cost of searching). We examined both participants’ decisions either to search or accept an offer and their RTs when they made these decisions. In short, we found first that participants planned ahead (S1A Fig, parameter “Prospective”). Second, when executing a sequence, participants were sensitive to the value and cost of searching (Fig 2A, parameters “Myopic,” “ProspVS1,” and “Cost”) and showed flexible adaptation of their strategy throughout the sequence (Fig 2A, parameter “ProspVS1-Adapted”), but also irrational biases that led them to search too many times. In particular they showed decision inertia, i.e., a “stuck in the rut” bias: The more participants had already searched, the more likely they were to search again, beyond what they should have done given the value of the alternatives (Fig 2A, parameter “# Prev Searches”). Participants also self-reported awareness of this bias (Fig 2B). Another bias was cost insensitivity during repeated searches, i.e., some participants did not take the repeatedly encountered costs sufficiently into account (Fig 2A, parameter “Cost”). Relatedly, participants reported “overchasing” of alternatives, i.e., searching too often because one of the alternatives appeared particularly appealing (Fig 2C). Third, we found that participants preemptively avoided trials on which a bias to search too many times would be particularly costly (S1A Fig, parameter “AvgProspVChange”). They self-reported awareness of this preemptive strategy (Fig 2D). For robustness of our modelling approach, see parameter recovery in Table A in S1 Text and S5 Fig (same data for real participants in Tables B and C in S1 Text). The results of the confirmation sample replicate almost all results from the discovery sample: For models of participants’ decisions (Fig 2A as well as S1A and S2A Figs), 13 out of 14 significant results replicated, i.e., the same factors that drove choices in the discovery sample were significant in the confirmation sample with the exception that in the discovery sample, accumulated cost during later searches on each trial reduced the probability of searching again, while it did not affect behaviour in the confirmation sample. For a description of those behavioural results, see the figure legend of Fig 2 as well as S1 and S2 Figs. For models of RT (S1B and S2B Figs), 10 out of 11 significant results replicated (exception: in the discovery sample, initial prospective value speeded up responses during later searches, while it did not impact behaviour in the confirmation sample).

Fig 2. Group-level results (confirmation sample).

Fig 2

(A) After initiating a search sequence, participants made a series of decisions of either banking an offer (and ending the trial) or searching again. We can assess, using our computational model (methods section “Analysis—Decisions”), whether participants behave mostly rationally and also whether they have decision inertia, i.e., show a “stuck in the rut” bias or a cost insensitivity bias. As in the discovery sample and as they should rationally, we found participants were more likely to go on searching if the myopic value was high (mean = 0.26, 95% CI [0.26 to 0.26]), if the initial prospective value (“ProspVS1”; i.e., the prospective value at the first search on each trial) was high (mean = 0.07, 95% CI [0.07 to 0.07]) or if the cost of searching was low (mean = −0.08, 95% CI [−0.08 to −0.08]) or if the drawn offers were low (mean = −0.49, 95% CI [−0.49 to −0.48]). Participants also adjust their estimate of prospective value (“ProspVS1-Adapted”) as it decreased over time as fewer and fewer available searches remained in a trial (mean = −0.03, 95% CI [−0.04 to −0.03]). We also found evidence for decision inertia (“stuck in a rut” bias, “# Prev. searches”): The more participants had already searched, the more likely they were to search again (mean = 0.04, 95% CI [0.03 to 0.04]). We tested participants sunk cost fallacy, i.e., whether they were more likely to continue searching the more costs they had already paid on a trial. However, this was not the case (“TotalCost” is not positive, (mean = −0.00, 95% CI [−0.00 to −0.00]). (B) Despite the strong behavioural effects of decision inertia, about half of the participants reported having no such bias (self-report Q 1). (C) Contrary to this, about 80% of people reported having been biased towards overchasing a rewarding option at least somewhat (self-report Q2). (D) About half of participants reported preemptively avoiding trials when they were worried they might spin the wheel too many times (self-report Q3). Error bars show Bayesian 95% credible intervals (2 tailed), significance is shown by credible intervals not including zero. Data in panel A are available in S9 Data (raw data in S1 Data) and panels B to D in S2 Data.

Clinical profile and dimensions

Participants were not specifically selected to have a minimum level of psychiatric symptoms (Table 1). Nevertheless, a sizable proportion of self-reported symptoms were in a clinically significant range using standard clinical cutoffs (Fig 3A, Table D in S1 Text), similar to previous findings in online samples [36,37]. For example, 30% of participants scored as having moderate or severe depression, and 43% of participants scored as having significant OCD symptoms.

Fig 3. Psychiatric symptom questionnaires in the confirmation sample.

Fig 3

(A) Histograms of the distribution of total questionnaire scores for the OCI-R (Ai), the AMI (Aii), the BDI (Aiii), and the Liebowitz social anxiety scale (Aiv). On each histogram, cutoffs based on previously published normative data are highlighted. (B) Correlation matrix (Pearson’s r) for all subscales included in the factor analysis. Squares highlight the factors subscales were assigned to in the factor analysis. (C) Factor loadings (i.e., the contribution of each subscale to each factor) of the factor analysis of the subscales. Highlighted in black are loadings above 0.4, for ease of visualisation. Data in panel A are available in S5 Data, panel B in S3 Data, and panel C in S16 Data. AMI, Apathy Motivation Index; BDI, Beck Depression Inventory; OCI-R, Obsessive-Compulsive Inventory.

We derived transdiagnostic dimensions cutting across 8 questionnaires (23 subscales, Fig 3B). Specifically, in the discovery sample (S3 Fig), we performed an exploratory factor analysis on the standard subscale summary scores extracted from each questionnaire. This resulted in 4 factors.

Our 4 factors could be labelled “compulsivity,” “apathy,” “depression/anxiety,” and “social anxiety.” The compulsivity factor mainly comprised subscales from the Obsessive-Compulsive Inventory (OCI-R) questionnaire (with a notable exception of the “obsessing” subscale) and the subscale measuring “unusual experiences” (e.g., hearing voices or telepathy) from the schizotypy questionnaire. The apathy factor included 2 of the 3 subscales from the Apathy Motivation Index (AMI): social and emotional apathy, but not the third subscale, behavioural apathy; it also included the anhedonia subscale of the schizotypy questionnaire and the “lack of importance of emotions/externally oriented thinking” from the Toronto Alexithymia questionnaire (including items such as “being in touch with emotions is essential”).

As the factor analysis used oblique rotation, the factor scores between dimensions were somewhat correlated (Table 2, highest correlation r = 0.65 between social anxiety and depression/anxiety, all others r < 0.38).

Table 2. Correlation between factor scores (***p < 0.001) for the confirmation sample based on factor loadings from either exploratory factor analysis done on the confirmation sample (“[C]”) or based exploratory factor analysis from the discovery sample (“[D]”).

DA[C] SA[C] C[C] A[C] DA[D] SA[D] C[D] A[D]
DeprAnx (DA)[C] 0.60*** 0.36*** 0.23*** 1.00*** 0.62*** 0.38*** 0.14***
SocialAnx (SA)[C] 0.60*** 0.30*** 0.25*** 0.63*** 0.99*** 0.34*** 0.16***
Compulsivity (C)[C] 0.36*** 0.30*** 0.03 0.38*** 0.32*** 0.99*** -0.06
Apathy (A)[C] 0.23*** 0.25*** 0.03 0.27*** 0.26*** 0.05 0.96***
DeprAnx[D] 1.00*** 0.63*** 0.38*** 0.27*** 0.65*** 0.40*** 0.17***
SocialAnx[D] 0.62*** 0.99*** 0.32*** 0.26*** 0.65*** 0.36*** 0.17***
Compulsivity[D] 0.38*** 0.34*** 0.99*** 0.05 0.40*** 0.36*** -0.05
Apathy[D] 0.14*** 0.16*** -0.06 0.96*** 0.17*** 0.17*** -0.05
Computed correlation used Pearson method with listwise deletion

Both solutions were virtually identical (correlations between e.g., Apathy[C] and Apathy[D]; all r > 0.96). Neither compulsivity nor apathy were correlated to any of the other factors at more than r > 0.38 (i.e., no shared variance r2>14.4%). Nonetheless, in all analyses, we controlled for all transdiagnostic dimensions. Data shown in this table are available in S16 Data.

Validation of factor analysis

We validated the factor analysis in the confirmation sample using confirmatory and exploratory factor analysis.

First, we ran a confirmatory factor analysis. Our comparative fit index (CFI) was 0.815, which indicates moderate, but not great fit [68]. However, our standardised root mean square residual (SRMR) and root mean square error of approximation (RMSEA) was relatively high at 0.097 and 0.112, respectively, indicating a relatively poor fit. This could be because of differences in how the exploratory and confirmatory factor analysis extracted dimensions. For example, the confirmatory factor analysis required a categorial assignment of each questionnaire subscale to a specific factor, while our questionnaire subscales could sometimes load onto more than one dimension (Fig 3C, S3C Fig). Therefore, we checked reliability of our factor solution further using exploratory factor analysis.

Specifically, we ran an exploratory factor analysis on the confirmation sample, resulting in 4 factors, as in the discovery sample. Visual inspection of the loadings of subscales to factors in the confirmation (Fig 3C) and discovery sample (S3 Fig) showed that both were virtually identical. To quantify this, we computed the factor scores for each person (e.g., a person’s “compulsivity score”) in the confirmation sample for each factor based either on factor loadings (i.e., how much each subscales contributes to a factor) from the exploratory factor analysis of the discovery sample or of the confirmation sample. Confirming the impression from the similarity of the factor weights in both analysis, these were extremely highly correlated (r between 0.96 and 1; Table 2).

Therefore, we next proceeded with our preregistered analysis strategy of applying the factor weights from the discovery sample to the confirmation sample to obtain factor scores for each person and factor. We then related these transdiagnostic dimensions to behavioural measures.

Testing of preregistered hypotheses in the confirmation sample

Next, we tested the relationship between the transdiagnostic dimensions of apathy and compulsivity with the behavioural and self-report measures described above. We simultaneously controlled for demographic measures (age, gender, and education), as well as psychiatric factors of no interest (depression/anxiety and social anxiety) and general measures of behavioural stochasticity in the sequential decision-making task (e.g., inverse temperature, as a measure of choice noisiness). For RT analyses, we controlled for average RT and RT variability. We only report results of preregistered analyses here. All of these were significant in the discovery sample. All preregistered tests and hypotheses are listed in the methods. Additional control analyses including medication status as additional control variable can be found in Table E in S1 Text.

However, due to a between subject indexing error in the RT analysis in the discovery sample, the results reporting relationships between RT effects and transdiagnostic dimensions were invalid. We therefore did not attempt to repeat these analyses. None of the other analyses were affected, and the decision data and self-report data by themselves were sufficient to confirm and disconfirm our 3 main hypotheses. In each case, the hypotheses had been framed in terms of decision effects (i.e., in terms of which choices were taken as opposed to how quickly a choice was taken) and differing self-reported insights. The RT effects had only served as secondary evidence suggesting RTs were sensitive to the same factors as decisions.

Results of testing hypothesis 1: Apathy is linked to decision inertia

Our first hypothesis was that apathy symptoms relate to increased decision inertia, i.e., being “stuck in a rut.” We measured this using a computational model that gave us for each participant a measure of how likely to search again depending on how often they had already searched, beyond what they should do given the value, cost, and number of opportunities (measured as in Fig 2A, parameter estimate for “# Prev Searches,” see also Fig 1F). Indeed, we found that apathy increased decision inertia (Fig 4A, mean = 0.066, 1-sided 95% CIlower: 0.005, hypothesis 1A; for all regression analyses relating behaviour to clinical factors, see S4 Fig for corresponding correlation analyses). In other words, participants with more apathy symptoms were more likely to search again the more they had already searched. To understand this effect in more detail, we wanted to exclude other possible explanations for the decision inertia (S5 Fig) by linking measures of several of these alternative effects to apathy. First, participants might be overoptimistic about how likely they were to receive a desired outcome (Fig 2A, parameter “ProspVS1-Adapted”; see also S1 Text “Carrying out a choice sequence” for details of how this was computed in our model). We found no evidence in the choice data that apathy led to reduced downward updating of subjective estimates of prospective value as it objectively declined during the course of each sequence (mean = 0.014, 95% CI [−0.062 to 0.089])). Second, we also excluded the possibility that accumulated costs (or “sunk cost fallacy”) could drive their behaviour (Fig 2A, parameter “Total cos”; see also S1 Text for details of computation). Finally, we also considered whether apathy might relate to not taking costs into account while going through a sequence of searches (Fig 2A, parameter “Cost”). We found no evidence for this in their choices (mean = 0.004, 95% CI [−0.072 to 0.083]).

Fig 4. Apathy symptoms and habitual overpersevering (confirmation sample).

Fig 4

To test whether higher scores on the apathy factor related to decision inertia (“stuck in a rut” bias), we did a regression analysis predicting choices of the bias based on the psychiatric factor scores, controlling for age, gender, education, and general performance on the task. The behavioural measures were derived from the analyses in Fig 2A, i.e., each person’s computational model weight for the impact of the previous number of searches already done in a trial on choice (on the group level making people more likely to repeat a choice). (A) Higher apathy scores related to increased decision inertia, i.e., the higher the apathy score, the more participants were likely to search again the more they had already searched. Error bars show 95% Bayesian credible intervals (1 tailed). Significance is shown by credible intervals not including 0 and dark blue colour. Data in this figure are available in S9 and S16 Data.

Results of testing hypothesis 2: Compulsivity and oversearching due to cost insensitivity

Our second hypothesis was that compulsivity would also relate to searching too often once participants started a sequence of searches, but that it would do so via a different mechanism than the one operating in apathy. Specifically, we hypothesised that it would be due to reduced sensitivity to costs (Fig 2A, parameter “Cost”; see also S1 Text “Carrying out a choice sequence” for model details). This had been the case in the discovery sample: Higher compulsivity was linked to taking costs into account less. In the confirmation sample, there was a similar effect but it fell just short of the threshold for significance (Fig 5A, mean = 0.061, 1-sided 95% % CIlower: −0.006, hypothesis 2A). However, the effect was of a similar size in both discovery and confirmation samples; notably, the estimated mean of the confirmation sample was within the 95% credible interval of the discovery sample [38,39]. In other words, the mean of the confirmation sample lay within the range of expected population means that had been established in the discovery sample, suggesting a similar pattern of behaviour exists in both discovery and confirmation samples.

Fig 5. Compulsivity scores, cost insensitivity during sequences of choices, and preemptive avoidance of biases (confirmation sample).

Fig 5

During later searches, higher compulsivity scores related to insensitivity to costs (A). Compulsivity was further related to increased self-reported chasing (B). (C) Increased compulsivity scores were, however, also related to increased self-reported preemptive avoidance of trials, which participants perceived as likely to lead to oversearching. (D) A mediation analysis revealed that the cost insensitivity bias during a sequence of choices was related (full mediation) to self-reported preemptive avoidance via self-reported awareness of a tendency to overchase. Stats in A–C are 95% Bayesian credible intervals (1 tailed); significance is shown by credible intervals not including zero and dark blue colour. D reports results of frequentist t tests at p < 0.05. Data in this figure are available in S2, S9, and S16 Data.

In addition to this somewhat ambiguous result, we had a complementary self-report measure of “overchasing” (Q2), i.e., an inability to stop searching because the alternatives appear attractive (even in the face of costs that lead to search termination in other participants). Here, we found that higher compulsivity scores were consistently related to more self-reported “overchasing” (Fig 5B, odds ratio = 1.224, 1-sided 95% % CIlower:1.081, hypothesis 2B). Also see exploratory results for an association between self-reported cost insensitivity and compulsivity, based on an additional self-report measure included in the confirmation sample.

Results of testing hypothesis 3: Compulsivity is linked to attempts to preemptively avoid situations in which sequential biases could arise

One rational response to knowing that one is biased when carrying out a sequence of searches would be to preemptively avoid situations where that bias is particularly likely to occur. We tested whether higher compulsivity scores were related to this preemptive avoidance and found this to be the case in terms of self-reported preemptive avoidance (Q3). We observed that compulsivity was linked to self-reported preemptive avoidance (Fig 4D, odds ratio = 1.423, 1-sided 95% % CIlower:1.255, hypothesis 3C). One possible interpretation is that these 2 behaviours are linked by awareness of the bias; in other words, people consciously avoid situations that can lead to the emergence of their latent bias if they are aware of the bias. We tested this using a mediation analysis. We found that, indeed, awareness of “chasing” mediated the relationship between reduced cost sensitivity during a sequence of searches and self-reported preemptive avoidance (Fig 4E, indirect path: cost insensitivity [later searches] to overchasing [self-report] B = 0.18, z = 5.06, p < 0.001 and overchasing [self-report] to preemptive avoidance [self-report] B = 0.32, z = 9.09, p < 0.001; but cost insensitivity [later searches] not directly to preemptive avoidance [self-report] B = −0.05, z = −1.29, p = 0.20, hypothesis 3C).

Additional control analyses

We collected medication status in the confirmation sample (Table 1) to exclude it as another potential confound and found no change in the clinical associations reported above (Table E in S1 Text).

Exploratory analyses—Predicting real-life individual differences through task behaviour and self-report

In addition to our prespecified tests, we wanted to know whether our cognitive task and self-reports of task behaviour, as a whole, had predictive power over the clinical and nonclinical demographic measures (Table 3). For this, we tested how well individual differences in the confirmation sample (in e.g., compulsivity) could be predicted from a model linking these measures to task/self-report measures in the discovery sample (see Exploratory Methods for details).

Table 3. Predicting transdiagnostic dimensions and demographics (confirmation sample).

Clinical/demographic factor Significance in confirmation sample and correlation/neg LL Behavioural measure Discovery sample mean and Bayesian 95% CI (2 tailed) Confirmation sample mean and Bayesian 95% CI (1 tailed)
Apathy N/A PrevSearch [later dec] 0.13 (0.03 to 0.23) 0.07 (0.01 to 0.13)
Compulsivity p < 0.0001,
r = 0.195
Cost [later dec] 0.15 (0.05 to 0.26) 0.06 (−0.01 to 0.13)
Offer [later RT] −0.11 (−0.21 to −0.01) 0.02 (−0.05 to 0.09)
Props Val S1 [later RT] 0.10 (0.00 to 0.20) −0.02 (−0.09 to 0.05)
Myopic [later RT] 0.13 (0.02 to 0.24) −0.03 (−0.10 to 0.04)
invTemp [initial dec] −0.1 (−0.21 to −0.00) −0.21 (−0.27 to −0.14)
Preempt Avoid [self rep] Odds ratio 1.46 (1.20 to 1.79) Odds ratio 1.42, CI90% (1.26 to 1.61)
Chasing [self rep] Odds ratio 1.32 (1.10 to 1.59) Odds ratio 1.22 (1.08 to 1.38)
DeprAnx p = 0.45,
r = 0.005
DecrThres [self rep] Odds ratio 1.35 (1.09 to 1.69) N/A
Offer [later RT] 0.13 (0.02 to 0.23) −0.02 (−0.09 to 0.05)
Cost [later RT] 0.10 (0.03 to 0.18) −0.06 (−0.13 to 0.02)
Myopic [later RT] −0.11 (−0.21 to −0.01) 0.03 (−0.05 to 0.10)
SocialAnx p = 0.25,
r = 0.027
RT.main [later RT] 0.12 (0.01 to 0.23) −0.07 (−0.15 to 0.01)
RT.main [initial RT] 0.13 (0.01 to 0.25) 0.01 (−0.07 to 0.10)
Age p = 0.0034
r = 0.104
Cost [later dec] 0.11 (0.01 to 0.21) 0.09 (0.02 to 0.15)
CostXPrevSearch [later dec] 0.11(0.01 to 0.21) 0.08 (0.01 to 0.14)
Offer [initial dec] −0.12 (−0.23 to −0.02) −0.13 (−0.19 to −0.06)
Myopic [initial dec] 0.10 (0.01 to 0.20) 0.12 (0.05 to 0.18)
Forage bias [initial dec] −0.12 (−0.22 to −0.02) −0.06 (−0.13 to 0.00)
Cost [later RT] −0.13 (−0.24 to −0.03) 0.02 (−0.05 to 0.08)
RT.main [later RT] 0.34 (0.25 to 0.43) 0.03 (−0.03 to 0.09)
RT.noise [later RT] −0.16 (−0.27 to −0.05) −0.07 (−0.13 to 0.01)
Forage Choice [later RT] 0.15 (0.04to 0.26) 0.03 (−0.03 to 0.10)
Preempt Avoid [initial RT] 0.16 (0.05 to 0.27) −0.03 (−0.09 to 0.04)
PreemptAvoid [self rep] Odds ratio 0.82 (0.67 to 0.99) Odds ratio 0.79 (0.70 to 0.89)
Gender p = 0.0253
negLL = 499
CostXPrevSearch [later dec] 0.23 (0.03 to 0.44) 0.15 (0.01 to 0.28)
RT.main [later RT] −0.46 (−0.65 to −0.27) −0.05 (−0.18 to 0.08)
RT.noise [later RT] 0.39 (0.19 to 0.59) 0.15 (0.02 to 0.28)
Inv temp [initial Dec] 0.39 (0.18 to 0.60) 0.25 (0.12 to 0.38)
RT.noise [initial RT] 0.45 (0.25 to 0.64) 0.03 (−0.10 to 0.17)
Offer [initial RT] 0.27 (0.10 to 0.45) 0.05 (−0.06 to 0.17)
Cost [initial RT] 0.15 (0.01 to 0.28) 0.04 (−0.07 to 0.16)
Myopic [initial RT] −0.23 (−0.40 to −0.05) −0.05 (−0.17 to 0.07)
Plan ahead [self rep] Odds ratio 1.75 (1.13 to 2.69) N/A
Education p = 0.0353
r = 0.065
ProspValDiff [later dec] −0.13 (−0.26 to −0.03) −0.09 (−0.16 to −0.02)
Forage Bias [later dec] 0.17 (0.02 to 0.30) 0.00 (−0.07 to 0.08)
RT.noise [later RT] 0.14 (0.03 to 0.29) 0.05 (−0.02 to 0.12)

Regression models fitted on the discovery sample data and predicting transdiagnostic and demographic measures were applied to the confirmation sample. Significance of the predictions was assessed either using correlations between predictions and measured data (all variables other than gender) or using “neg LL” as measure of goodness of model fit (gender). In summary, compulsivity was the factor we could predict best from a mixture of self-report and behaviour. While we could also predict, age, gender, and education, the included behavioural/self-report measures are listed separately for the discovery sample (95% 2-tailed Bayesian CI, based on analyses predicting the behavioural/self-report measures on all clinical and demographic measures) and the confirmation sample (95% 1-tailed CIs); measures that replicated are highlighted in grey. The self-report factor DecrThres (“Would you accept offers you would have previously rejected?”) was not part of any preregistered hypotheses and thus accidently not collected in the confirmation sample, which is why it could not be used in the confirmation sample. See Tables F–H in S1 Text for full list of all links between clinical/demographics measures and task behaviour and self-reports. Data in this table are available in S2, S9S12, and S16 Data.

CI, credible interval; neg LL, negative log likelihood; RT, reaction time.

We found that compulsivity, age, gender and education, but not depression/anxiety or social anxiety, could be significantly predicted based on task behaviour and self-reports (Table 3, column “Significance in confirmation sample and correlation/neg LL”).

Apathy was only associated with a single significant measure in the discovery sample, described above, and therefore was not tested again here.

We noted that predictions based mainly on RT measures were least likely to replicate (Table 3, column “Confirmation sample mean and Bayesian 95% CI (1 tailed)”).

Exploratory analyses—Measuring cost insensitivity through self-report

We also collected a direct self-report question of cost insensitivity (S2D Fig), on top of the related over chasing and task behaviour measure, to see whether this is also linked to compulsivity. As we expected, we found a strong link between the 2 (odds ratio = 1.292, 95% 2-sided CI [1.117 to 1.497]), further supporting our findings of compulsivity being related to changes in how costs are taken into account.

Discussion

Summary

In natural environments, decisions are not made in isolation from one another. Instead for many active decision-making agents, whether human or animal, the choices taken at one point in time may have implications for which opportunities may be encountered later [7,10,13]. We were interested in how participants flexibly plan and adjust sequences of decisions and how this relates to different psychiatrically relevant dimensions in a large confirmation sample of participants (n = 756). We preregistered analyses based on an initial discovery sample (n = 449) and found a high degree of reproducibility, i.e., 4 out of the 5 tests supporting our 3 hypotheses fully replicated. Specifically, we found that both apathy and compulsivity related to failures to stop searching appropriately, but for different reasons. Apathy was linked to “decision inertia,” i.e., a tendency to search again and again for a new offer, especially when the participant had searched several times before, regardless of whether this was appropriate. In contrast, compulsivity was linked to changes in how people thought about costs; while the evidence of increased insensitivity to costs in choices was somewhat ambiguous, there was clear evidence of changes in the perception of costs both when directly asked about cost insensitivity and as a sense of “overchasing,” i.e., the feeling of going after high value options longer than they should. Individuals who scored highly on compulsivity also reported taking preemptive action to avoid situations in which their bias would have affected their decision-making. Our results suggest that apathy and compulsivity are linked to distinct biases in sequential search.

Transdiagnostic approach, ecological paradigm, and preregistration

Traditionally, psychiatric research has thought of clinical disorders as defined by a set of specific symptoms. However, more recently, researchers have begun to look at transdiagnostic dimensions of symptoms, cutting through traditional disorder boundaries [15]. In this framework, dimensions such as apathy or compulsivity emerge. Importantly, these dimensions exist on a spectrum from health to disease.

Teasing apart the role of distinct yet partially correlated transdiagnostic symptoms was made possible here by testing a large number of participants over the internet. For this approach to work well, it is necessary that symptom scores span the range from healthy to psychiatrically relevant. We found here, as have others previously [37,40,41], that participants in an internet based study showed just such a range of symptoms and in fact even showed symptoms at higher rates than found in the normal population. To ensure that the self-reported symptoms were valid, we included several quality checks. Previous work [20] that directly compared transdiagnostic dimensions derived from self-reports over the internet to clinician rated diagnoses found good correspondence between them and crucially also that dimensions related more clearly to behaviour than diagnostic labels.

We used a sequential search task as many decisions in real life involve planning a sequence of decisions and carrying them out while monitoring whether the preplanned strategy is still appropriate. Such decisions are often made between a current offer and the opportunity to carry on searching for even better offers. This contrasts with more commonly studied tasks involving single decisions between specific well-defined offers. We and others have shown that these processes are neurally and behaviourally distinct [4,8,42]. We used a previously established computational model [14] to tease apart the different component cognitive processes underlying search behaviour, and in this way, we identified different types of sequential bias. We complemented this approach with analysis of participants’ self-reports of their own behaviour to test their metacognition. We ensured good data quality by giving participants automated training on the task and testing their knowledge using a multiple choice test they had to pass before being admitted to the main study. As a consequence, data quality was very high (only few participants had to be excluded based on task performance).

Many of our results highlight the usefulness of acquiring reports of subjective experience related to the task performed (see also [43]): Participants themselves might not have direct access to the decision model (including biases) they implement. Instead, they make observations about their own behaviour. Understanding such metacognition of decision-making preferences and biases further complements our understanding of the cognitive processes during decisions based on cognitive models of the task behaviour itself. If we want to positively affect people’s behaviours, understanding how people think about what they are doing could be of great benefit. In particular, understanding the interactions of cognitive models with experience, behavioural strategies, and metacognition, as we have done here, could help the development of better cognitive approaches to improve behaviours long term.

In this study, we took a 2-step discovery and confirmation sample approach to combine statistical rigour with the possibility that the data themselves might help inform our hypotheses. We preregistered the study based on the results of the discovery sample. The ability to collect very large samples of participants over the internet with ease allowed us to achieve high power (95%) even for associations between clinical factors and task behaviour of a small effect size (correlations ~ r = 0.1). Overall, we found that the evidence relating to our 3 primary hypotheses was reproducible. In the future, it might be worthwhile considering performing power calculations assuming a smaller effect size than found in a discovery sample [39].

Apathy

The dimension we termed “apathy” pulled together measures from different questionnaires related to emotional and social apathy, anhedonia and externally oriented thinking (from alexithymia questionnaire [44], example item: “I prefer just to let things happen rather than to understand why it turned out that way”) in agreement with previous studies that have highlighted strong conceptual and empirical overlap between the constructs of apathy and anhedonia [19]. Of note, we found that the apathy dimension was distinct from another dimension capturing other symptoms of depression, anxiety, and fatigue, in agreement with previous findings in neurodegenerative disorders [45], Parkinson disease [46], depression [47], and healthy volunteers [35]. In contrast, however, behavioural apathy appeared distinct and was related to the depression/anxiety dimensions. This suggests that apathy is multifaceted.

We found that the higher participants scored on our dimension of apathy the more decision inertia they showed. Our results highlight how apathy can counterintuitively lead to the exertion of more unnecessary effort (i.e., searching too often). Unnecessary effort exertion could potentially lead to further exhaustion and apathy and, therefore, as a result of a vicious cycle, to the exertion of more and more unnecessary effort. Within work psychology, “overcommitment,” defined as an inability to disengage from work has been highlighted as an intrinsic factor modulating the impact of the working environment [24,48,49]. Our findings are also in broad agreement with a previous report in a small sample of patients with depression, which is often comorbid with apathy, of delayed stopping in the “secretary problem” task [50]. We also note that we did not find an effect of apathy on general response slowing at the initial decision. This was potentially the case because our task had a fixed “monitoring” period (3 to 6 seconds) on each trial to ensure participants considered all information before responding. However, this may have reduced sensitivity to individual differences in RT.

Our findings may initially seem at odds with other reports that apathy or anhedonia lead to avoidance of effort when participants choose between 2 options that vary in effort and reward [5153]. However, importantly, participants make a very different kind of decision in our paradigm—they make sequences of decisions. This means that at each choice point, previous decision plans can be retrieved and potentially adjusted, i.e., the decision can be reconsidered. Thus, thinking about stopping, is in itself, a deliberate cognitive act, which is affected by apathy. This interpretation would be in agreement with previous clinical reports that patients with apathy show normal behaviour if externally prompted, but struggle to self-generate behaviour [22], as well as experimental findings of reduced generation of new options as candidates for decision-making [23].

Compulsivity

The dimension we called “compulsivity” included different subscales of the OCI-R [54], with the notable exception of the obsessive thinking subscale, as well as the unusual experience subscale of the schizotypy questionnaire (short scales for Measuring Schizotypy [55]). We did not include broader questions measuring compulsivity in the context of other disorders such as eating disorders or gambling. Compulsivity was primarily related to changes in self-report.

Compulsivity led to self-reported “overchasing”, i.e., the subjective experiences of searching for too long in the hope of getting a better offer, which, in turn, was linked to a reported tendency to preemptively avoid even beginning to explore potentially costly courses of action. This interpretation was bolstered by a finding of clear changes in self-reported cost insensitivity in an exploratory analysis. However, we found that objective measurements of cost sensitivity per se were less clearly related to compulsivity. While they were correlated with compulsivity in the discovery sample, such an association was not quite significant in the confirmation sample. However, on the other hand, it is equally important to note that the confirmation sample mean fell within the Bayesian credible interval of possible population means of the discovery sample; such a finding is often taken as evidence of confirmation of a result [38,39]. Importantly, using a higher inclusion threshold for the effect in the discovery sample, such as p < 0.005 as has been suggested [56], would not have changed this result, as the original effect was highly significant.

Our self-report changes of overchasing and cost insensitivity are also in agreement with the clinical profile of OCD patients carrying out compulsions despite their immense time costs [28]. It is also related to experimental findings of reduced cost sensitivity in the context of oversampling information before making a decision [57,58]. Our findings are also in broad agreement with previous reports linking compulsivity to an increase of model-free (or “habitual”) mechanisms controlling behaviour rather than model-based (or “goal-directed”) mechanisms [5961]. What our findings suggest is that the reason that control might change from model-based to model-free is because of an underestimation of the costs associated with the model-free behaviour (i.e., in our case the sequential search costs). While the most commonly used psychological treatment, cognitive behavioural therapy is based on the assumption that obsessions are primary [25], this has been challenged. Specifically, while there are patients with compulsions without obsessions [29], the reverse is rarely the case [30]. If obsessions were primary to compulsions, one would expect compulsions only to emerge after specific obsessions have developed. Our findings align more with the view that compulsions can be a primary deficit. This is in agreement with the ego-dystonic nature of OCD and recent findings of intact metacognition in OCD, but failure to use this knowledge to behave appropriately [62]. However, whether OCD leaves metacognitive abilities completely intact is unclear; a study using a transdiagnostic, large-scale approach, instead of a case–control approach, suggested compulsivity can affect both metacognition and behaviour [40]. In fact, worse metacognitive ability and overconfidence in one’s ability have also been found in other paradigms [63,64], suggesting that metacognitive processes and behaviour are not completely independent.

Finally, our data suggest that people perceive themselves to be implementing preemptive avoidance strategies, i.e., try to avoid situations where they fear they will behave suboptimally in the future, as a function of their compulsive symptoms. Importantly, we found that the cost insensitivity bias did not directly drive the appearance of a preemptive strategy. Instead, the subjective experience of overchasing for better options (despite the costs being incurred) fully mediated the relationship between the cost insensitivity bias and self-reported preemptive behaviour. These findings align with clinical reports that OCD patients avoid situations that provoke their symptoms by, e.g., asking family members to wash their clothes for them [32]. Of particular note in our data is that such avoidance can be related to the behavioural bias per se, rather than a specific obsession. This is again in contrast to a view that obsessions are primary in OCD and that patients avoid situations because they trigger their obsessions, which then, in turn, trigger their compulsions.

While we find that compulsivity is linked to increased self-reported preemptive avoidance, there was no such link to a corresponding behavioural measure aimed at this (Table G in S1 Text). This could suggest either that participants only subjectively perceive a change in their behaviour, which is in fact unchanged, or it could suggest that the behavioural measure does not capture the full range of behaviours participants might engage in order to preemptively avoid but which they consider when they self-report.

Predicting individual differences with behaviour and self-reports

In exploratory analyses, we used a data mining/prediction approach to build models that used all potentially informative measures together to predict participants’ transdiagnostic and demographic features similarly to [65]. These models were built based on the discovery sample, taking advantage of all significant associations in the data. We tested predictive performance in the confirmation sample. We found that compulsivity, age, gender, and education could be predicted from a combination of task behaviour and self-reports. In contrast, depression/anxiety and social anxiety could not be predicted. Note that apathy was not included in this analysis as there was only one significant association in the discovery sample, i.e., the link with decision inertia described above, meaning no model using a combination of measures for its prediction could improve predictive power.

Interestingly, within this exploratory analysis, we noticed that the least replicable links between task behaviour and clinical/demographic features were related to RTs. This could have been because our online setup was not optimised for RT collection, e.g., we did not impose response deadlines or reward speeded responding.

Overall, the analysis highlights the possibility that links between metacognition, task behaviour, and clinical dimensions can be multifaceted, i.e., not map one to one with each other. In our data, this was the case for compulsivity, which was best predicted by a mixture of metacognitive measures (e.g., self-reported chasing) and behavioural measures (such as choice stochasticity). Importantly, to find and validate such multifaceted relationships, a discovery confirmation sample approach is particularly valuable as it allows for the initial identification of the combination of the relevant factors in the discovery sample and a test of these factors in the confirmation sample. While we had no prior hypotheses about age and gender, the mixtures of effects are in line with previous studies showing differences in risk taking between male and female participants [65] and changes in decision-making with age [66]. However, again, they highlight how with complex traits, a larger set of behavioural and metacognitive changes can emerge in more complex tasks.

Summary

In conclusion, during complex search sequences, behavioural biases emerged in a way that was related to participants’ tendencies towards compulsivity and apathy. Combining our clinical dimensions with our computational model and self-reports of participants’ own task experiences, we could further specify how participants’ behaviour, experience, and behavioural strategies interact and how those relationships link to differences in their clinical profile.

Methods

Deviations from the Stage 1 protocol are indicated by superscript letters; these deviations can be found in the section “Protocol deviations.” To enable both data-informed hypothesis generation and testing as well as statistical rigour, we followed a 2-step discovery and confirmation sample procedure. First, we collected a discovery sample, which we analysed and used to define a precise set of hypotheses linking clinical dimensions with our task behaviour. This was accepted as a Stage 1 Preregistered Research Article. See the OSF link here: https://osf.io/dfg2u. The results of the discovery sample together with the preregistered hypotheses are all contained in the preregistration. Hypotheses of interest that were preregistered concerned relationships between clinical questionnaires and task behaviour. These are 1 tailed. For any statistical tests establishing group average effects of behaviour, we used 2-tailed tests as they were not included in the preregistered hypotheses.a

Participants

Participants were recruited via the online platform Prolific.ac. We aimed to acquire 750 data sets that fulfilled the preregistered inclusion criteria detailed below; 756 were collected instead due to technical reasons in the Prolific system. This sample size was determined based on the discovery sample shown below to give a power of 0.95 (see Fig 6 and section “Power calculation” below for details). Ethical approval for the study was given by the Oxford University Central University Research Ethics Committee (CUREC) (Ref-number: R54722/RE001). All participants provided informed consent before taking part in the study.

Fig 6. Power analysis.

Fig 6

Power increases with sample size. To reach 95% power in the analysis with lowest power (i.e., analysis of the impact of apathy on overperseverance measured in the participants’ decisions), we selected 750 participants for the confirmation sample.n RT, reaction time.

Task

In a computerised foraging task, participants made repeated choices with the aim to earn as much money as possible (Fig 1C). On each of 150 trials, participants were shown one option that we refer to as the “default offer” (point range 5 to 150 in steps of 5). The offer was the default option that the participants would have unless they decided to spin a “wheel of fortune.” The alternatives on the wheel differed in their magnitudes (same range as initial offer) and in their probability (i.e., the size of the area they occupied on the wheel). If participants decided to “bank” the offer, the trial was finished. If, however, they decided to spin the wheel, they first had to pay a cost (either 0, 7, or 15 points, fixed for the whole trial, but varied across trials) and then spun the wheel. They would get a new offer depending on where the wheel stopped. There were between 2 and 6 alternatives on the wheel on each trial. Participants were shown how often they could spin the wheel on each trial (“spins remaining”; varied between 1 and 10 times across trials); they had to pay the same cost for each spin.

Timings: There was a random delay of 3 to 6 seconds (uniform) between the start of a trial and when participants could make their choice (buttons were shown as “disabled”) to ensure that participants thought carefully about their choice. If participants chose to spin the wheel, the wheel spun for a random duration of 0.3 to 2 seconds. There was no enforced delay for any choices after the first one on each trial. When a trial was finished, there was a constant delay of 1.5 seconds before participants could move on to the next trial.

Training: Participants were given standardised instructions and 5 training trials. After, they had to complete a multiple choice test. If they passed the test, they could move on the task; if they failed it, they had to repeat the instructions and the multiple choice test.

Design of task schedule: We constructed 10 random task schedules that fulfilled the following criteria (see S1 Text, section “Task schedule design and model validation” and S7 Fig and Table A in S1 Text for details): First, correlations between variables influencing decisions (see analyses below for details) were below 0.7. Second, parameter recovery showed Pearson correlations of r > 0.85 between simulated and recovered parameters for all parameters. Third, parameter recovery across all schedules showed r > 0.85 for all parameters, suggesting that parameter estimates obtained using different schedules were comparable i.e., there was no systemic schedule bias affecting the results. Any given participant shares the schedule they performed with a tenth of the other participants.

The study was programmed in JavaScript and the study was run using JATOS [67].

Questionnaires

After the task, participants completed a selection of questionnaires to assess a range of psychiatric symptoms: AMI [68], OCI-R [54], Beck Depression Inventory (BDI) [69], Spielberger State Anxiety [70], Liebowitz Social anxiety [71], Short Scales for Measuring Schizotypy [55], (“unusual experiences” and “introvertive anhedonia” subscale), Fatigue scale [72], and Toronto alexithymia scale [44]. Before completing the questionnaires, participants were shown a message highlighting the importance of the questionnaires for the study and were told that some questions might repeat and asked to answer as consistently as possible. Five questions were repeated to screen for data quality (see below). If participants responded faster than the fastest quartile of participants from the discovery sample with included data, they were shown a warning and asked whether they wanted to check their answers again.b

We also included several questions for participants to report their metacognition (i.e., insight) into how they behaved during the task. As question related to insight about decision inertia, we asked participants, “Did you ever feel like you continued spinning the wheel just because you had spun the wheel already this round, rather than because you really thought this was the right thing to do?” (Q1) For the question related to overchasing, we asked, “Did you ever feel like you spun too many times in a round because there was an option on the wheel that you really wanted to get?” (Q2) For the question related to preemptive avoidance, we asked, “Did you ever avoid starting spinning the wheel in the first place in a round because you were worried that you might end up spinning it too many times?” (Q3). All self-reports were presented as a scale from “never” to “always” in 7 steps.

To obtain additional control covariates for analyses, we also asked participants about their age, gender, years of education (General Certificate of Secondary Education [GCSE] A-levels/High school, BSc, MSc, and PhD) and whether they were currently taking any psychoactive medication or drugs (which will be coded as distinct categories whenever drugs listed by at least 10 participants fell in a category, e.g., serotonergic, noradrenergic, dopaminergic, GABA-ergic, and “other”; otherwise just listed as “psychoactive medication yes/no”) and, if applicable, frequency and dose. It turned out that in the confirmation sample, responses to specific medications were too infrequent and therefore medication was coded as “yes/no.”

Participant exclusion criteria

As our data were acquired online, we took several steps to ensure data quality. First, we ensured participants fully understood the task instruction by giving them an automated tutorial followed by a multiple choice test that they needed to pass with 100% correct before being able to do the task. Second, we gave participants bonus payments depending on their performance to increase motivation. Third, we excluded participants based on task performance as well as poor completion of the questionnaires. We excluded participants whose behaviour either suggested they did not engage with the task or whose responses were too unbalanced to allow analysis. Behavioural analyses were performed separately for the initial search decision on each trial and for all later searches, see below, and data exclusion reflected this. Participants were excluded due to bad performance if, when considering only the initial choice of each trial, for the 10 trials with highest offer minus search value (i.e., strongest value pressure to accept) they accepted the initial offer less often than for the 10 trials with the lowest offer minus search value. Participants were excluded due to imbalanced responses if—separately for the initial search decision of each trial and for later searches—they had fewer than 25 accept or spin responses.

Exclusion based on questionnaire responses:

  1. Not completing all psychiatric questionnaires

  2. Five questions were repeated. If a participant’s responses were different by 2 or more points on average, then they were excluded.

Participants who only failed inclusion criteria for the behavioural models were included in the factor analyses of the questionnaire data (see below). Participants who only failed inclusion criteria for one behavioural model (e.g., only for the initial search) were included in the analyses of the questionnaire data and the other behavioural model.

Data analysis

Data analysis was performed in MATLAB and R (version 4.0.4), including the packages rstan [73] (version 2.21.2), shinystan [74] (version 2.5.0), brms [7578] (version 2.15.0), psych [79] (version 2.1.3), processRc (based on PROCESS [80], version 0.2.6) implemented using lavaan [81] (version 0.6–8), ggplot2 [82] (version 3.3.3), ggpubr [83] (version 0.4.0), tidyverse [84] (version 1.3.1), and sjPlot [85] (version 2.8.7).

Computational model of prospective value

We used a previously described decision tree model to determine the value of searching for each trial [14]. In short, this model took into account the possibility of searching more than once on a given trial (up to the maximum number of searches available), the magnitudes and probabilities of the alternatives and the cost of searching. See S1 Text (“Decision tree model to derive prospective value”) for details. We illustrate the key factors contributing to our model-based prospective value estimate (see Fig 1D). For this, we first conducted a regression predicting prospective value based on the average value of all alternative options in the environment (myopic value), the standard deviation of all alternatives, the number of searches available, the cost of searching, and all possible 2- to 4-way interactions. To compute how well these components together approach prospective value, we computed the variance explained (R2).

Analysis—Decisions

We analysed choices to accept an offer or search with a decision model. In the model, we predicted decisions based on a linear combination of the different factors that should drive choices. We analysed the initial search decision on each trial and the subsequent searches separately. This means that we allow for differences in the processes underlying the decisions. This may be the case because for the initial decision, participants had time (forced delay of 3 to 6 seconds) to monitor the options and plan ahead (and we found previously that indeed they planned ahead [14]). In contrast, later decision could be executed immediately and could have been preplanned prior to their onset.

In the model of the initial search, we included as factors as previously [14]: myopic value (i.e., the average of the alternatives), prospective value (see above), the offer value, the cost of searching, and how much prospective value might change over the course of a trial. The last factor captures a measure for each person to what extent they preemptively avoided trials in which sequential bias would be more costly (as the faster prospective value might decrease, the more costly it could be to oversearch). All individual effect sizes of factors influencing participants’ decisions can subsequently be related to clinical dimensions and introspective measures.

Specifically, we first computed the utility of searching versus accepting the offer (“Decision model—first search”):

Utility=bSearchBias+bOffer*Offer+bCost*Cost+bMyopic*Myopic+bProspectve*Prospective+bAvgProspVChange*Averageprospectivevaluechange,

where parameters (e.g., bSearchBias) were allowed to vary for each person (see below). For computational reasons [86], parameters were standardised by the some of their absolutes:

k=abs(bSearchBiasPre)+abs(1)+abs(bCostPre)+abs(bMyopicPre)+abs(bProspectivePre)+abs(bAvgProspVChangePre)

and then as an example,

bSearchBias=bSearchBiasPrek

The variable “average change in prospective value” above captures how much the prospective value might change over the course of a trial, meaning that participants would need to adapt their behaviour more.

AvgProspValChange=ProspValSearch1ProspVallastsearchbeforeProspcVal=0Maximumnumberofsearches

The utility was then translated into a probability of choosing to search using a standard soft-max decision rule:

p(Search)=1εβ*Utility,

where β was the inverse temperature, i.e., a measure of choice noisiness.

Note that this decision model is a reparameterised logistic regression analysis [14] with the main difference being that one single parameter (inverse temperature) is used to capture choice noisiness, rather than choice noisiness affecting the size of the estimates of all parameters (e.g., if a participant is noisy, in a regression all regression weights might appear smaller than for a less noisy participant, thus introducing strong correlations between the estimated parameters which hinders the goal of relating specific parameters, i.e., behavioural measures, to specific psychiatric factors). It also means all regression weights (bs) are relative weights, i.e., represent the relative contribution of that regressor on the decision.

The model was fitted as a hierarchical model using full Bayesian inference with MCMC sampling [73]; all variables were z-score normalised across all participants to ease fitting. In the hierarchical model, we specified weak priors for all group level parameters (normal distribution with mean 0 and standard deviation 3). The parameters for individual participants were then related to the group level distribution as [parameteri ~ normal(meangroup,stdevgroup)). For all parameter other than the inverse temperature, inspection of individual participants’ parameters of the discovery sample (from a nonhierarchical version of the model) suggested that a normal distribution on the parameters was appropriate, while for the inverse temperature, a normal distribution was applied to log transformed data (log(invTempi) ~ normal(log(invTempgroup),stdevgroup). Models were run for 4,000 samples per chain (of which 2,000 were warmup samples)d, with 4 chains. Absence of divergent samples was verified using the shinystan diagnostic toolbox [74] and convergence of the model fit was assessed using the Rhat <1.1 criterion [87]. In the confirmation sample, significance of regression weights will be assessed using 1-tailed (2-tailed in the discovery sample) 95% Bayesian credible intervals.

We analysed the later search decisions on each trial to obtain different measures of sequential biases. To capture decision inertia, we included the previous number of searches on a trial. To capture sunk cost fallacy, we included the total costs incurred so far on a trial. To capture whether participants did not adapt their impression of how valuable it was to search on a given trial as fewer and fewer searches remained, we included how much the prospective value had changed since the start of the trial. To capture whether participants ignored the cost of searching in favour of chasing high valued option, we included the cost of searching (“Decision model—later searches”):

Utility=bSearchBias+bOffer*Offer+bCost*Cost+bMyopic*Myopic+bProspectiveVS1*ProspectiveValueSearch1+bProspVS1Adapted*(ProspectiveValueSearch1CurrentProspectiveValue)+bPreviousSearches*Previousnumberofsearches+bTotalCost*TotalCost,

where again the regression weights were regularised as described above, by dividing through k, i.e., the sum of their absolutes.e

We checked how well the behaviour of the fitted model resembled the behaviour of real participants (posterior predictive checks). For this, we extracted, for each participant and for each search decision, the model’s probability of searching as opposed to accepting the offer, given the parameters of the best fitting model. These are shown in Fig 1E and 1F together with the data from real participants.

Analysing behaviour—RTs

We analysed how different factors influenced participants’ RTs using multiple regressions similar to the decision analyses to look at the factors influencing deliberation times. For the initial decision, we can look at which planning factors participants deliberate over during the initial planning. In the later decisions, we can look at the factors participants consider during execution and updating of the decision strategy as the number of opportunities depletes.

First, as there was no response deadline, outlier RTs (very slow RTs) were possible. RTs over 10 seconds were therefore discarded. After outlier rejection, RTs were modelled using a hierarchical regression with a shifted lognormal distribution, fit using the brms package that implements Bayesian multilevel models in R, using exactly the same regressors as for the decision model (“RT model—initial search”), shown here in standard regression modelling notation (as implemented in R):

RT1+Offer+Cost+Myopic+Prospective+Averageprospectivevaluechange+(1+Offer+Cost+Mypoic+Prospective+Averageprospectiivevaluechange|ID),

where “|ID” appears in Wilkinson notation and indicates that all regressors were given a “random” in addition to a “fixed” effect or, in other words, that they were modelled hierarchically. We also allowed decision-to-decision variability of the RTs to vary between participants (following a hierarchical distribution). All regressors were z-score normalised. We used the standard priors in brms for the group level (i.e., flat prior for group level regression weights for all parameters). Convergence was again assessed as described above. Models were run for 12,000 samples (of which 8,000 warmup samples) with 3 chains and within chain parallelisation.f

For the later searches (“RT model—later searches”), we again included exactly the same regressors as for the corresponding model of decisions.

RT1+Offer+Cost+Myopic+ProspectiveValueSearch1+(ProspectiveValueSearch1CurrentProspectiveValue)+Previousnumberofsearches+TotalCost+(1+Offer+Cost+Myopic+ProspectoveValueSearch1+(ProspectiveValueSearch1CurrentProspectiveValue)+Previousnumberofsearches+TotalCostID)

Factor analysis of questionnaire measures

Summary: In the discovery sample, we performed an exploratory factor analysis. Here, this was validated using confirmatory and exploratory factor analysis. As this showed good replication of results, we next applied the factor weights from the discovery sample to the confirmation sample to obtain factor scores for each person for each factor.g

In detail: We performed exploratory factor analysis of the discovery sample to derive the transdiagnostic clinical dimensions within the psychiatric questionnaires using the R package “psych.” The aim of factor analysis is to explain the relationship (correlation) between the questionnaire items or subscales as a function of underlying factors. For this, we computed the previously established subscale summary scores for each questionnaire and entered those scores into a factor analysis. We used summary scores rather than raw questionnaire responses because, first they have better properties (e.g., they have better reliability and better distribution shapes and are continuous rather than categorical, allowing Pearson correlations to be computed) [88]. Second, they offer a better ratio of number of participants to subscales than to items. We followed standard procedures with the factor analysis [89]: First, we obtained the correlation matrix of all subscales (Spearman) as this is the data that factor analysis aims to explain. The second step was to determine how many factors were appropriate to describe our data. For this, we used parallel analysis [90]. Third, we performed a factor analysis, using generalised least square to fit the model. The final step involved “rotating” the factors (in a 2D coordinate system, one could visualise moving the 2 axes). This is done to improve the interpretability of the factors—unrotated, the results of factors analysis are such that items often show a strong relationship with several factors (high “crossloading”). After rotation, this is reduced, making the meaning of each factor easier to interpret, but allowing correlation between the factors. We used oblique rotation (oblimin) for this step. This revealed 4 factors in the discovery sample that we labelled “Apathy," “Compulsivity,” “Depression/Anxiety,” and “Social anxiety.”

Validation of factor analysish

To validate the factor analysis in the confirmation sample, we ran 2 complementary checks. First, we performed a confirmatory factor analysis. For this, we used the Lavaan package [85], specifying 4 factors, as in the exploratory factor analysis on the discovery sample. Factors were allowed to correlate but subscales were assigned to a single factor (the one with the highest loading in the discovery sample). However, in reality, the loadings of some subscales were actually relatively evenly spread across factors, making the categorical assignment less than ideal (Fig 3, S3 Fig). We judged fit using 3 distinct indices, i.e., CFI, SRMR, and RMSEA.

As the results of the confirmatory factor analysis did not unambiguously suggest that the 4-factor solution was a good fit to the data, we followed our prespecified procedure and next performed an exploratory factor analysis on the confirmation sample. The same methods were followed as described above. We then visually compared the results of the factor analyses of discovery and confirmation sample. To quantify the similarity between the factor scores (for each person, for each factor) obtained if either weights of the discovery or of the confirmation sample were applied to the confirmation sample, we correlated them.

Relating questionnaires to task performance and self-reports

We used the discovery sample to establish the exact links to be tested in the confirmation sample between questionnaire measurements, i.e., clinical dimensions and task performance (decisions, RTs, and post-task self-reports). Specifically, for the analyses of decisions and RTs, we performed linear regressions (R using Stan and brms software) in which the behavioural/self-report measures were related to the 4 psychiatric factors (compulsivity, apathy, depression, and social anxiety), while controlling for the covariates gender (coded as categorical variable with levels female, male, or “other”); education (coded as ordered factor [75] with levels: GCSE, A-levels/High school, BSc, MSc, and PhD; and age. In the analyses of parameters derived from the decision model, we included inverse temperature as a measure of general task performance and choice consistency. Similarly, in the analysis with parameters of the RT model, we included the intercept (as measure of general RT) and sigma (as measure of RT variability) from the model described above as covariate. In the analyses of the post-task self-reports, responses were treated as ordinal and monotonically increasing (rather than continuous) and therefore modelled with a cumulative distribution as implemented in brms [76]. All continuous regressors were z-score normalised. To illustrate the relationship between psychiatric factors and behaviour, we also plot partial correlations of the residuals. Residuals were computed in 2 regressions. First, the behaviour of interest was predicted as a function of all factors (i.e., psychological and control measures) listed above, apart from the one of interest (e.g., apathy). Second, the residual variance in the behaviour was correlated with the questionnaire dimension of (e.g., apathy) using nonparametric correlations (Kendall’s tau) to ensure results were not driven by outliers. For the analyses of the self-report questionnaire scores, residuals could only be computed by treating them as continuous.

The exact same procedure was applied to both discovery and confirmation sample. Significance in the confirmation sample was assessed using 1-tailed 95% Bayesian credible intervals (1 tailed, as the hypotheses were derived from the discovery sample). For the linear regressions weights (predicting behaviour based on clinical dimensions), Bayesian credible intervals excluding zero were judged as significant and confirming our hypotheses, while credible intervals including zero were judged as not significant. For the ordinal regressions (predicting self-reports based on task behaviour), Bayesian credible intervals on the parameter estimates (odds ratios) were judged as significant if they did not include one and they were judged as not significant if they included one.i We choose the 95% credible intervals as those are the intervals that determine the 95% likelihood of containing the true parameter value, given our data.

Mediation and moderation analyses (see below) were run using the R package processR (via R package lavaan) [80,91]. Significance was be determined as p < 0.05 1 tailed.

From the discovery sample, we had derived specific compound hypotheses that were be tested in the confirmation sample (summarised in Table 3). For completeness, we also report all significant findings from the discovery sample again in the confirmation sample but marked as exploratory (see below).

We have removed all elements of the 3 hypotheses that were linked to RTs as an indexing error in the discovery sample that was used to define statistical test meant those were incorrect. We include the removed text explaining the RT elements in the footnotes below and show it in Table 4 as greyed out.j

Table 4. Summary of the hypotheses and tests.

Research question Hypothesis Statistical test Outcome measure Power Outcome
H1 Apathy is linked to decision inertia H1A Apathy is linked to greater decision inertia Regression weight relating apathy to b_#PrevSearches (from the analysis “Decision model—later searches”) Significant: 95% 1-tailed Bayesian credible interval excludes zero; Not significant: 95% credible interval includes zero 0.950 Confirmed
H1B Apathy is linked to greater response speeding with the number of searches already done Regression weight relating apathy to b_#PrevSearches (from the analysis “RT model—later searches”) 0.998 Not tested
H2 Compulsivity is linked to a bias to oversearch due to cost insensitivity H2A Compulsivity is linked to reduced sensitivity to costs encountered during the later searches (i.e., after the initial search) Regression weight relating compulsivity to b_Cost (from the analysis “Decision model—later searches”) 0.982 Not confirmed
H2B Compulsivity is linked to increased self-reported chasing Regression weight relating compulsivity to self-reported chasing (self-report Q2) 0.996 Confirmed
H2C Compulsivity is linked to increased slowing with costs encountered during the later searches Regression weight relating compulsivity to b_Cost (from the analysis “RT model—later searches”) 0.996 Not tested
H3 Compulsivity is linked to preemptive avoidance H3A Compulsivity is linked to increased self-reported preemptive avoidance Regression weight relating compulsivity to self-reported preemptive avoidance (self-report Q3) 0.996 Confirmed
H3B Relationship between cost insensitivity and preemptive avoidance is mediated by metacognitive awareness Mediation analysis of cost insensitivity (b_Cost from “Decision model—later searches”) to self-reported preemptive avoidance (self-report Q3) by self-reported chasing (self-report Q2) Sig.: p-value of mediation <0.05 (1 sided); not significant: p-value > 0.05 1.000 Confirmed

All statistical tests consisted of a multiple linear regression, predicting the behavioural measure of interest based on the clinical factors of interest while controlling for all other clinical factors, as well as for age, gender, and education. Power is indicated for a sample size of 750 participants. See Methods—“Relating questionnaires to task performance and self-reports” and Fig 5 for further details. Elements of hypotheses that were based on an indexing error in the RT analysis of the discovery sample are greyed out for clarity and were not tested in the confirmation sample. The last column indicates which hypotheses were confirmed, partially confirmed or not confirmed in the confirmation sample.m

Hypothesis 1: Apathy is linked to decision inertia

To test this hypothesis, we used as dependent variable the measure of decision inertia from the decision model of the later searches (i.e., each person’s parameter b#PrevSearches). In the discovery sample, the higher the apathy, the more participants tended to search again the more they have already searched (i.e., larger b#PrevSearches)—hypothesis 1A.k

Hypothesis 2: Compulsivity is linked to a bias to oversearch due to cost insensitivity

To test this hypothesis, we used as dependent variable the measure of cost sensitivity from the decision model of the later searches (i.e., each person’s regression weight bCost). In the discovery sample, the higher the compulsivity score, the more a participant was insensitive to costs—hypothesis 2A. In the discovery sample, participants were aware of their bias, and, therefore, compulsivity was also linked to increased self-reported over chasing (self-report Q2)—hypothesis 2B.l

Hypothesis 3: Compulsivity is linked to preemptive avoidance of bias sensitive situations

To test this hypothesis, we used as dependent variable the self-reported preemptive avoidance (self-report Q3). In the discovery sample, higher compulsivity was linked to more preemptive avoidance (hypothesis 3A). Finally, we tested directly the relationship between the different behaviours and self-reports affected by compulsivity. In the discovery sample, the relationship between the behavioural measure cost insensitivity during later searches (regression weight bCost) and self-reported preemptive avoidance (self-report Q3) was mediated by metacognitive awareness, i.e., the self-reported over chasing (self-report Q2)—hypothesis 3B.

Power analysis

To determine the sample size required for the confirmation sample, we performed a Bayesian power analysis for each of the Bayesian analyses listed above and a non-Bayesian power calculation for the mediation analysis in hypothesis 3B (Fig 6). Sample sizes were selected to give at least 95% power for each analysis. For the Bayesian power analysis, we repeatedly simulated samples of data given the parameter estimates from our discovery sample (respecting the covariance between the various psychological and demographic factors) and checked the percentage of samples where the Bayesian credible interval excluded zero (i.e., we calculated the power).

A summary of all differences between discovery and confirmation sample can be found in Table 5.

Table 5. Differences between the discovery and confirmation sample.

Discovery sample Confirmation sample
Sample size (after exclusions) 449 756
Number of questionnaires included 13, mean: completion time: 25 minutes (AMI, AUDIT, BDI, BIS, FAS, FFMQ, IDS, OCI, PANAS, STAI, SPQ, TAS, and LSAS) 7, completion time: 17 minutes (AMI, BDI, FAS, OCI, SPQ, TAS, and LSAS)
Questionnaire delivery After behavioural task without further instruction After behavioural task, instructions about importance for study and warning that questions will be repeated to check consistency of answers; warning given for each questionnaire if time to complete is lower than expected; 5 questions are repeated (“check question”)
Exclusion criteria for questionnaire (1) Not completing all questionnaires; (2) Scoring opposite extreme scores on questions with similar wording across questionnaires; (3) Responding faster than expected reading time on at least 2 questionnaires; (4) Difference of less than 0.5 points between reversed and nonreversed items on at least 2 questionnaires (1) Not completing all questionnaires; (2) Difference of more than 1 point on at least 2 repeated questions (out of 5)
Number of trials in task 200 (45 minutes to complete) 150 (36 minutes to complete)
Statistics of relationship between psychological factors and task behaviour/reports of behaviour Two-tailed One-tailed for preregistered analyses; 2 tailed for exploratory analyses
Factor analysis technique to derive factor scores Exploratory factor analysis Application of factor weights derived from discovery sample
Construction of schedule 10 schedules designed to be sensitive to presence or absence of behavioural effects on population as a whole (see [14]) 10 schedules designed to give good parameter recovery (correlation of r > 0.8 between simulated and recovered parameters, within and between schedules)

AMI, Apathy Motivation Index; AUDIT, Alcohol Use Disorders Identification Test; BDI, Beck Depression Inventory; BIS, Barratt Impulsiveness Scale; FAS, Fatigue Assessment Scale; FFMQ, Five Facets of Mindfulness Questionnaire; IDS, Inventory for Depressive Symptomatology; LSAS, Liebowitz Social Anxiety Scale; OCI, Obsessive-Compulsive Inventory; PANAS, Positive Affect and Negative Affect Schedule; SPQ, Short scales for measuring schizotypy; STAI, State-Trait Anxiety Inventory; TAS, Toronto Alexithymia Scale.

Exploratory analysis

In addition to the preregistered analyses, we wanted to test whether our task behaviour and self-report measures, as a whole, had predictive power over the transdiagnostic dimensions and demographics. We did this by building and testing overall predictive models from the discovery sample as follows:

  • Step 1: In the discovery sample, we established which behavioural/post-task self-report measures showed significant links to the transdiagnostic/demographic measures based on the regression analyses described above (e.g., predicting behavioural measures based on all transdiagnostic and demographic measures). Our cutoff for including a regressor was whether the 95% Bayesian confidence interval included 0 or not.

  • Step 2: We then fit regression models to the discovery sample in which transdiagnostic/demographic measures were included as predicted variables and all behavioural/self-report measures identified in step 1 were included as predictors.

  • Step 3: We applied the models from step 2 to the confirmation sample. This thus generated predictions for each person in the confirmation sample of their transdiagnostic/demographic measures.

  • Step 4: We compared these predicted measures to the actually observed transdiagnostic/demographic measures to check the goodness of the predictions. For all continuous measures, this was done by correlating predicted and measured transdiagnostics/demographics. For the binary measure (gender), we used the negative log likelihood instead. Significance was assessed using permutation tests (10,000 permutations; randomly shuffled predictions).

In future studies, such analyses could easily be prespecified (as has been done before [70]). However, even without preregistration this approach is appealing precisely because it reduces a potentially large set of regressor-to-factor relationships to a small number of sets that are each based on the discovery sample and validated on the confirmation sample, while limiting the number of tests to one per factor of interest.

Added question to questionnaire

Based on the discovery sample results and the resulting hypothesis 2, we added a self-report measurement directly targeting cost insensitivity (S2D Fig) in addition to overchasing. However, because it was not part of the discovery sample, we did not include it in the preregistered tests. The exact wording of the cost insensitivity question (self-report Q4) was “Did you ever feel like you spun too many times in a round because you did not take enough into account how expensive it was to spin?,” and we analysed it exactly like the other self-report measures.

Protocol deviations

a We added this summary after the preregistration to improve clarity.

b In the preregistration and the discovery sample, this was specified as “responded faster than the time it took one author (JS) to read the questions.” It was changed to the present procedure to be more objective. Unfortunately, the preregistration had not been updated to reflect this.

c In the preregistration processR was used, but was no longer compatible with the latest version of R. Version numbers have been added for all R packages.

d The number of samples was increased compared to the preregistration as more were needed to achieve convergence using the prespecified criteria.

e We cut the following text from the preregistration “As previously [14], we also tested (using R toolbox processR) whether the relationship between decision inertia (i.e., regression weight b#PreviousSearches) and preemptive avoidance (i.e., regression weight bAvgProspVChange) was moderated by awareness of decision inertia (self-report Q1). Significance will be tested at p < 0.05 1 tailed in the confirmation sample (and 2 tailed in the discovery sample),” as we did not run those analyses because they were not essential for any of our hypotheses.

f The number of samples was increased compared to the preregistration as more were needed to achieve convergence using the prespecified criteria.

g This summary was added after the preregistration to increase clarity.

h The level of detail specifying the confirmatory factor analysis was insufficient. We therefore removed the old short version and added this section instead. Importantly, all steps were carried out as prespecified, i.e., we did not diverge from our analysis plan.

i This explanation was added for clarity but the analysis was exactly the same procedure that we had already used for the discovery sample.

j This explanation was added after the preregistration as we discovered an indexing error in the discovery sample during the analysis of the confirmation sample.

k We removed the following text and analysis “And we will include the corresponding measure from the RT model of the later searches—hypothesis 1B. We predict that the higher the apathy, the more participants will speed up with having done more searches,” as the discovery sample RT analysis was erroneous.

l As in H1, we removed the following text and analysis related to RT “Participants are aware of the cost insensitivity and they will try (but fail) to counteract it as it occurs, slowing them down. We predict therefore that compulsivity will be linked to a greater slowing of RT with increased cost (i.e., each person’s regression weight for bCost from the RT analyses)—hypothesis 2C.”

m The added this explanation to the table after the preregistration as we found the indexing error during the analysis of the confirmation sample.

n This figure still contains the RT based results for determining power as we used them when deciding on confirmation sample size.

Ethical approval plan

Ethical approval has already been obtained for the discovery and the confirmation sample (see above).

Supporting information

S1 Text

Contains Supporting information Methods, Results, and Supporting information Tables A–I. Table A: Parameter recovery (schedule used for confirmation sample). The table shows correlations (Pearson’s r) between simulated (“ground truth”) parameters (headers highlighted in grey) and the fitted (“recovered”) parameters (headers with white background) for the decision models of the initial (top) and later (bottom) searches. All parameters showed very good recovery (r > 0.88, diagonal, highlighted in blue). The off-diagonal values show that confusion between parameters was very low (all < 0.22). We used 500 simulated participants for these results. Data in files S13 and S14 Data. Table B: Parameter correlations for initial decisions (confirmation sample). We ran a Pearson correlation of the hierarchically fit parameter estimates from our cognitive model on the initial decisions to look at their relationships. Even though simulations (S1 Table) showed very low parameter confusions, parameters from real participants showed in parts substantial correlations (Pearson’s r), suggesting those correlations are a real relationship existing in participants, rather than an artifact of model fitting. invTemp (inverse temperature of the softmax equation linking value to choice probabilities), SearchBias (bias for or against searching). Prospective (prospective value of the model), Myopic (Myopic value of the model). AvgProspVChange (average change of prospective value per search in the future). Data in file S11 Data. Table C: Parameter correlations for later searches (confirmation sample). We ran a Pearson correlation of the hierarchically fit parameter estimates from our cognitive model on the later decisions to look at their relationships. Even though simulations (S1 Table) showed very low parameter confusions, parameters from real participants showed in parts substantial correlations (Pearson’s r), suggesting those correlations are a real relationship existing in participants, rather than an artifact of model fitting. Abbreviations same as S2 Table. ProspVS1 (prospective value at first search). ProspVS1mAdapted (change of prospective value since first search). PrevSearches (number of previous searches). Total Cost (sum of total cost since first search). Data in file S9 Data. Table D: Additional demographics (confirmation sample). Participants completed the following questionnaires. AMI [1], OCI-R [2], BDI [3], Spielberger State Anxiety [4], Liebowitz Social anxiety [5], Short Scales for Measuring Schizotypy [6] (“unusual experiences” and “introvertive anhedonia” subscale), Fatigue scale [7], and Toronto alexithymia scale [8]. Depression cutoffs were (less than 13 “Minimal,” less then 20 “Mild,” less than 29 “Moderate,” and higher or equal 29 “Severe”). Obsessive compulsion less than 21 “none,” otherwise “clinically significant”). For social anxiety, we had less than 30 as “none,” between 30 but less than 60 as “possible” and at and above 60 as “probable” social anxiety. For Alexithymia, we had less than 52 as none, between 52 but less than 61 as “possible” and at and above 62 as “present” Alexithymia. Fatigue was “none” below 22, “present” between 22 and 34, and “extreme” above 34. Data in file S5 Data. Table E: Results of regressions linking clinical dimensions to decision and self-report measures controlling for medication status (confirmation sample). As control regressors, we included age, gender and education (as in main manuscript). Additionally, here, psychoactive medication status is also included (coded as “Yes”/“No” as there were too few respondents of individual types of medications). Data in files S2, S6, S9, and S16 Data. Table F: Results of regressions linking clinical dimensions to later search decisions. Results of all parameters from model fit to the later search decisions and dimensions. We used 90% 2-sided CIs, to show the results of 1-sided (95% CI) tests for our preregistered hypotheses and to show what other links have any evidence. Data in files S9 and 16 Data. Table G: Results of regressions linking clinical dimensions to initial search decisions. Results of all parameters from model fit to the initial search decisions and dimensions. We used 90% 2-sided CIs, to show the results of 1-sided tests for our preregistered hypotheses and to show what other links have any evidence. Data in files S11 and S16 Data. Table H: Results of regressions linking clinical dimensions to self-report measures. Results of all self-report measures fit to the dimensions. We used 90% 2-sided CI, to show the results of 1-sided tests for our preregistered hypotheses and to show what other links have any evidence. Data in files S2 and S16 Data. Table I: Additional medication information (confirmation sample). All information for the types of medications excluded and included participants used (No and Yes). The most common medication were selective serotonin reuptake inhibitor. Data in file S6 Data. AMI, Apathy Motivation Index; BDI, Beck Depression Inventory; CI, confidence interval; OCI-R, Obsessive-Compulsive Inventory.

(PDF)

S1 Fig. Effects on decisions and deliberation time during initial decisions (confirmation sample).

(A) Using a decision model, we tested which factors influenced the decision of whether to initiate a search. We found that participants were more likely to search when prospective and myopic values were high and that vice versa they were less likely to search when initial offer value or costs of searching were high. (B) We analysed the impact of different factors on RTs using a regression analysis. Of particular note, the larger the prospective value, the slower participants responded, potentially indicating more time taken to plan ahead. (C) When directly asked about their preemptive avoidance about half never reported using such strategies (self-report Q3). (E) Histogram of RTs across all participants, for initial searches on each trial (red) and later searches (blue). Note that for initial decisions there was a fixed delay of 3 to 6 seconds on each trial before participants could make responses, which is not included in the RT measured here. RTs for initial searches have an earlier mode than for later searches, but more slow RTs (“thicker tail” of the distribution). Error bars show Bayesian 95% credible intervals (2-tailed), significance is shown by credible intervals not including zero. Data of A in file 11, B in 12, C in 2, and D in 1. RT, reaction time.

(TIFF)

S2 Fig. Effects on decisions and deliberation time during later decision sequence (confirmation sample).

(A) Once participants engaged with a sequence of searches, they had to decide how long to go on for. We found evidence for decision inertia (“# Prev. searches”) that was independent of effects of “sunk cost” fallacy (“Total cost”) or lack of updating of prospective value, i.e., their search strategy (“ProspVS1-Adapated”). Across all participants, we found sensitivity to costs (“Cost”). (B) RTs for later searches revealed that the more often participants had searched already within a given sequence, the faster they got (“#Prev. searches”). Moreover, as in the initial search (S1B Fig), the higher the cost, the slower they responded on these later searches in the sequence (“Cost”). (C) Model fits for the decision data, higher values indicate worse model fit (sum of log likelihoods across all participants and all trials, computed using cross-validation, relative to the best fitting model). The “full model”—the model used throughout the manuscript—provides the best fit to the data. The other models were derived from the “full model,” leaving out individual components of the model, one at a time. (D) Participants were also asked about their own task behaviour, such as their perceived cost insensitivity and most reported insensitivity to the costs to some extent (question only included in the confirmation sample). (E) Despite the strong choice effects of decision inertia, about half of the participants reported having no such bias (self-report Q 1). (F) Contrary to this, about 80% of people reported having been biased towards overchasing a rewarding option at least somewhat (self-report Q2). Error bars show Bayesian 95% credible intervals (2-tailed), significance is shown by credible intervals not including zero. Data of A in file 9, B in 10, C in 15, and D to F in 2.

(TIFF)

S3 Fig. Psychiatric symptom questionnaire results in discovery sample.

(A) Histograms of the distribution of total questionnaire scores for the OCI-R (Ai), the AMI (Aii), the BDI (Aiii), and the Liebowitz social anxiety scale (Aiv). On each histogram, cutoffs based on previously published normative data are highlighted. (B) Correlation matrix (Pearson’s r) for all subscales included in the factor analysis. Squares highlight the factors subscales were assigned to in the factor analysis. (C) Loadings (i.e., the contribution of each subscale to each factor) of the factor analysis of the subscales. Highlighted in black are loadings above 0.4, for ease of visualisation. AMI, Apathy Motivation Index; BDI, Beck Depression Inventory; OCI-R, Obsessive-Compulsive Inventory.

(TIFF)

S4 Fig. Partial correlations between clinical factors and behavioural measures (confirmation sample), controlling for other clinical, demographic and general task performance measures (see Methods—“Relating questionnaires to task performance and self-reports”).

Significance was assessed using nonparametric correlations (Kendall’s tau). Panel A corresponds to regression coefficients shown in Fig 4. Panels B–D correspond to regression coefficients shown in Fig 5 A–C. Note that for correlations of clinical factors with self-report questionnaires (C and D), responses needed to be treated as continuous, while in the regression analysis they were treated as ordered factors. Data of A to D in file 9, 2, and 16.

(TIFF)

S5 Fig. Parameter recovery (confirmation sample).

We constructed 10 random schedules as detailed in the S1 Text section “Task schedule design and model validation.” Scatter plots show correlations between simulated and recovered parameters for the initial search (Ai) and later searches (Ai). Different colours show the different schedules (each with n = 500 simulated participants). Correlations (Pearson’s r) are shown across all 10 schedules. The average correlations (Pearson’s r) between regressors used in the RT regression and the decision-making models are at or below 0.7 for all variables for initial decisions (Bi) and later (Bii).” Data of A in file 13 and 14 and B in 1.

(TIFF)

S1 Data. Raw behavioural data (choices and RT) for each participant for each trial of the decision-making task.

RT, reaction time.

(CSV)

S2 Data. Raw questionnaire answers (for each item of each questionnaire) and post-task self-reports for each participant.

(CSV)

S3 Data. Summary scores for questionnaire subscales (separately for each questionnaire) for each participant.

See Fig 3B.

(CSV)

S4 Data. Demographics for each participant.

See Table 1.

(CSV)

S5 Data. Clinical questionnaire summed scores for each participant.

See Fig 3A and Table D in S1 Text.

(CSV)

S6 Data. Each participant’s medication status.

See Table E and Table I in S1 Text.

(CSV)

S7 Data. Each participant’s probability of searching binned by relative value.

See Fig 1E.

(CSV)

S8 Data. Each participant’s probability of searching binned by the number of previous searches and cost of searching.

See Fig 1F.

(CSV)

S9 Data. Computational decision model parameters for each participant for data fitted on the decisions during the later searches on each trial.

See Figs 2A, 4, and 5 as well as S2A and S4S6 Figs and Table 3.

(CSV)

S10 Data. Computational RT model parameters for each participant for data fitted on the decisions during the later searches on each trial.

See S2B Fig and Table 3. RT, reaction time.

(CSV)

S11 Data. Computational decision model parameters for each participant for data fitted on the decisions during the initial search on each trial.

See S1A Fig, Table 3, and Tables B and H in S1 Text.

(CSV)

S12 Data. Computational RT model parameters for each participant for data fitted on the decisions during the initial search on each trial.

See S1B Fig and Table 3. RT, reaction time.

(CSV)

S13 Data. Decision model simulation: simulated (“true”) and recovered (“fitted”) parameters for later searches of each trial.

See S5 Fig and Table A in S1 Text.

(CSV)

S14 Data. Decision model simulation: simulated (“true”) and recovered (“fitted”) parameters for the initial search of each trial.

See S5 Fig and Table A in S1 Text.

(CSV)

S15 Data. Model fits (cross-validation) for each participant for the later searches of each trial.

See S2C Fig.

(CSV)

S16 Data. Transdiagnostic dimensions (factor scores) for each participant.

See Figs 3C, 4, and 5, S4 Fig, Tables 2 and 3 as well as Tables E–I in S1 Text.

(CSV)

Abbreviations

AMI

Apathy Motivation Index

BDI

Beck Depression Inventory

CFI

comparative fit index

GCSE

General Certificate of Secondary Education

OCD

obsessive-compulsive disorder

OCI-R

Obsessive-Compulsive Inventory

RMSEA

root mean square error of approximation

RT

reaction time

SRMR

standardised root mean square residual

Data Availability

All anonymized data and analyses script are freely available on a repository (https://osf.io/dfg2u/). The data are also available in the supplementary data files (S1S16 Data).

Funding Statement

This work was funded by the Medical Research Council (MR/N014448/1, Skills Development Fellowship,JS; G0902373, grant MFSR, https://mrc.ukri.org/), the Biotechnology and Biological Sciences Research Council (BB/R01803/1, BB/V004999/1, AFL Fellowship NK; Discovery Fellowship JS, https://bbsrc.ukri.org/), the Clarendon Fund (studentship, HAT https://www.ox.ac.uk/clarendon) and the Wellcome Trust (WT100973AIA and 203139/Z/16/Z, https://wellcome.org/ MFSR). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Kolling N, Wittmann M, Rushworth MFS. Multiple Neural Mechanisms of Decision Making and Their Competition under Changing Risk Pressure. Neuron. 2014;81:1190–202. doi: 10.1016/j.neuron.2014.01.033 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.McNamara JM, Trimmer PC. Sequential choices using signal detection theory can reverse classical predictions. Behav Ecol. 2019;30:16–9. doi: 10.1093/beheco/ary132 [DOI] [Google Scholar]
  • 3.Korn CW, Bach DR. Maintaining Homeostasis by Decision-Making. PLoS Comput Biol. 2015;11:e1004301. doi: 10.1371/journal.pcbi.1004301 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Hayden BY, Pearson JM, Platt ML. Neuronal basis of sequential foraging decisions in a patchy environment. Nat Neurosci. 2011;14:933–9. doi: 10.1038/nn.2856 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Kolling N, Akam T. (Reinforcement?) Learning to forage optimally. Curr Opin Neurobiol. 2017;46:162–9. doi: 10.1016/j.conb.2017.08.008 [DOI] [PubMed] [Google Scholar]
  • 6.Kolling N, Behrens T, Wittmann M, Rushworth M. Multiple signals in anterior cingulate cortex. Curr Opin Neurobiol. 2016;37:36–43. doi: 10.1016/j.conb.2015.12.007 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Kolling N, Wittmann MK, Behrens TE, Boorman ED, Mars RB, Rushworth MF. Value, search, persistence and model updating in anterior cingulate cortex. Nat Neurosci. 2016;19:1280. doi: 10.1038/nn.4382 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Korn CW, Bach DR. Heuristic and optimal policy computations in the human brain during sequential decision-making. Nat Commun. 2018;9:1–15. doi: 10.1038/s41467-017-02088-w [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Korn CW, Bach DR. Minimizing threat via heuristic and optimal policies recruits hippocampus and medial prefrontal cortex. Nat Hum Behav. 2019;1. doi: 10.1038/s41562-019-0603-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Rushworth MF, Kolling N, Sallet J, Mars RB. Valuation and decision-making in frontal cortex: one or many serial or parallel systems? Curr Opin Neurobiol. 2012;22:946–55. doi: 10.1016/j.conb.2012.04.011 [DOI] [PubMed] [Google Scholar]
  • 11.Hayden BY. Economic choice: the foraging perspective. Curr Opin Behav Sci. 2018;24:1–6. doi: 10.1016/j.cobeha.2017.12.002 [DOI] [Google Scholar]
  • 12.Pearson JM, Watson KK, Platt ML. Decision Making: The Neuroethological Turn. Neuron. 2014;82:950–65. doi: 10.1016/j.neuron.2014.04.037 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Scholl J, Klein-Flügge M. Understanding psychiatric disorder by capturing ecologically relevant features of learning and decision-making. Behav Brain Res. 2018;355:56–75. doi: 10.1016/j.bbr.2017.09.050 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Kolling N, Scholl J, Chekroud A, Trier HA, Rushworth MFS. Prospection, Perseverance, and Insight in Sequential Behavior. Neuron. 2018;99:1069–1082.e7. doi: 10.1016/j.neuron.2018.08.018 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Gillan CM, Daw ND. Taking Psychiatry Research Online. Neuron. 2016;91:19–23. doi: 10.1016/j.neuron.2016.06.002 [DOI] [PubMed] [Google Scholar]
  • 16.Stephan KE, Bach DR, Fletcher PC, Flint J, Frank MJ, Friston KJ, et al. Charting the landscape of priority problems in psychiatry, part 1: classification and diagnosis. Lancet Psychiatry. 2016;3:77–83. doi: 10.1016/S2215-0366(15)00361-2 [DOI] [PubMed] [Google Scholar]
  • 17.Cuthbert BN. The RDoC framework: facilitating transition from ICD/DSM to dimensional approaches that integrate neuroscience and psychopathology. World Psychiatry. 2014;13:28–35. doi: 10.1002/wps.20087 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Ruscio AM, Stein DJ, Chiu WT, Kessler RC. The epidemiology of obsessive-compulsive disorder in the National Comorbidity Survey Replication. Mol Psychiatry. 2010;15:53–63. doi: 10.1038/mp.2008.94 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Husain M, Roiser JP. Neuroscience of apathy and anhedonia: a transdiagnostic approach. Nat Rev Neurosci. 2018;19:470–84. doi: 10.1038/s41583-018-0029-9 [DOI] [PubMed] [Google Scholar]
  • 20.Gillan CM, Kalanthroff E, Evans M, Weingarden HM, Jacoby RJ, Gershkovich M, et al. Comparison of the Association Between Goal-Directed Planning and Self-reported Compulsivity vs Obsessive-Compulsive Disorder Diagnosis. JAMA Psychiat. 2020;77:77–85. doi: 10.1001/jamapsychiatry.2019.2998 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Koutsimani P, Montgomery A, Georganta K. The Relationship Between Burnout, Depression, and Anxiety: A Systematic Review and Meta-Analysis. Front Psychol. 2019;10. doi: 10.3389/fpsyg.2019.00010 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Foussias G, Remington G. Negative Symptoms in Schizophrenia: Avolition and Occam’s Razor. Schizophr Bull. 2010;36:359–69. doi: 10.1093/schbul/sbn094 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Sinha N, Manohar S, Husain M. Impulsivity and apathy in Parkinson’s disease. J Neuropsychol. 2013;7:255–83. doi: 10.1111/jnp.12013 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Siegrist J, Li J. Associations of Extrinsic and Intrinsic Components of Work Stress with Health: A Systematic Review of Evidence on the Effort-Reward Imbalance Model. Int J Environ Res Public Health. 2016;13:432. doi: 10.3390/ijerph13040432 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Whittal ML, McLean PD. CBT for OCD: The rationale, protocol, and challenges. Cogn Behav Pract. 1999;6:383–96. doi: 10.1016/S1077-7229(99)80057-1 [DOI] [Google Scholar]
  • 26.Cassin SE, von Ranson KM. Personality and eating disorders: A decade in review. Clin Psychol Rev. 2005;25:895–916. doi: 10.1016/j.cpr.2005.04.012 [DOI] [PubMed] [Google Scholar]
  • 27.Figee M, Pattij T, Willuhn I, Luigjes J, van den Brink W, Goudriaan A, et al. Compulsivity in obsessive–compulsive disorder and addictions. Eur Neuropsychopharmacol. 2016;26:856–68. doi: 10.1016/j.euroneuro.2015.12.003 [DOI] [PubMed] [Google Scholar]
  • 28.Association AP. Diagnostic and statistical manual of mental disorders (DSM-5®). American Psychiatric Pub. 2013. doi: 10.1097/01.pra.0000430502.78833.06 [DOI] [PubMed] [Google Scholar]
  • 29.Mansueto CS, Keuler DJ. Tic or compulsion? It’s Tourettic OCD Behavior Modification. 2005;29:784–99. doi: 10.1177/0145445505279261 [DOI] [PubMed] [Google Scholar]
  • 30.Williams MT, Farris SG, Turkheimer E, Pinto A, Ozanick K, Franklin ME, et al. Myth of the pure obsessional type in obsessive–compulsive disorder. Depress Anxiety. 2011;28:495–500. doi: 10.1002/da.20820 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Gillan CM, Robbins TW. Goal-directed learning and obsessive-compulsive disorder. Philosophical Transactions of the Royal Society B: Biological Sciences. 2014;369:20130475–5. doi: 10.1098/rstb.2013.0475 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Albert U, Baffa A, Maina G. Family accommodation in adult obsessive–compulsive disorder: clinical perspectives. Psychol Res Behav Manag. 2017;10:293. doi: 10.2147/PRBM.S124359 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Anderson M, Magruder J. Split-Sample Strategies for Avoiding False Discoveries. Cambridge, MA: National Bureau of Economic Research; 2017. Jun p. w23544. Report No.: w23544. doi: 10.3386/w23544 [DOI] [Google Scholar]
  • 34.Fafchamps M, Labonne J. Using Split Samples to Improve Inference on Causal Effects. Polit Anal. 2017;25:465–82. doi: 10.1017/pan.2017.22 [DOI] [Google Scholar]
  • 35.Petitet P, Scholl J, Attaallah B, Drew D, Manohar S, Husain M. The relationship between apathy and impulsivity in large population samples. Sci Rep. 2021;11:4830. doi: 10.1038/s41598-021-84364-w [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Shapiro DN, Chandler J, Mueller PA. Using Mechanical Turk to Study Clinical Populations: Clinical Psychological Science. 2013. [cited 29 Jul 2019]. doi: 10.1177/2167702612469015 [DOI] [Google Scholar]
  • 37.Arditte KA, Çek D, Shaw AM, Timpano KR. The importance of assessing clinical phenomena in Mechanical Turk research. Psychol Assess. 2016;28:684–91. doi: 10.1037/pas0000217 [DOI] [PubMed] [Google Scholar]
  • 38.Patil P, Peng RD, Leek JT. What Should Researchers Expect When They Replicate Studies? A Statistical View of Replicability in Psychological Science. Perspect Psychol Sci. 2016;11:539–44. doi: 10.1177/1745691616646366 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Camerer CF, Dreber A, Holzmeister F, Ho T-H, Huber J, Johannesson M, et al. Evaluating the replicability of social science experiments in Nature and Science between 2010 and 2015. Nat Hum Behav. 2018;2:637–44. doi: 10.1038/s41562-018-0399-z [DOI] [PubMed] [Google Scholar]
  • 40.Seow TXF, Gillan CM. Transdiagnostic Phenotyping Reveals a Host of Metacognitive Deficits Implicated in Compulsivity. Sci Rep. 2020;10:1–11. doi: 10.1038/s41598-019-56847-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Ophir Y, Sisso I, Asterhan CSC, Tikochinski R, Reichart R. The Turker Blues: Hidden Factors Behind Increased Depression Rates Among Amazon’s Mechanical Turkers. Clin Psychol Sci. 2020;8:65–83. doi: 10.1177/2167702619865973 [DOI] [Google Scholar]
  • 42.Kolling N, Behrens TE, Mars RB, Rushworth MF. Neural mechanisms of foraging. Science. 2012;336:95–8. doi: 10.1126/science.1216930 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Rigato J, Rennie SM, Mainen ZF. The overlooked ubiquity of first-person experience in the cognitive sciences. Synthese. 2019. [cited 16 Oct 2019]. doi: 10.1007/s11229-019-02136-6 [DOI] [Google Scholar]
  • 44.Bagby RM, Parker JD, Taylor GJ. The twenty-item Toronto Alexithymia Scale—I. Item selection and cross-validation of the factor structure. J Psychosom Res. 1994;38:23–32. doi: 10.1016/0022-3999(94)90005-1 [DOI] [PubMed] [Google Scholar]
  • 45.Levy ML, Cummings JL, Fairbanks LA, Masterman D, Miller BL, Craig AH, et al. Apathy Is Not Depression JNP. 1998;10:314–9. doi: 10.1176/jnp.10.3.314 [DOI] [PubMed] [Google Scholar]
  • 46.Kirsch-Darrow L, Fernandez HF, Marsiske M, Okun MS, Bowers D. Dissociating apathy and depression in Parkinson disease. Neurology. 2006;67:33–8. doi: 10.1212/01.wnl.0000230572.07791.22 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.Marin RS, Firinciogullari S, Biedrzycki RC. Group differences in the relationship between apathy and depression. J Nerv Ment Dis. 1994. doi: 10.1097/00005053-199404000-00008 [DOI] [PubMed] [Google Scholar]
  • 48.Eddy P, Heckenberg R, Wertheim EH, Kent S, Wright BJ. A systematic review and meta-analysis of the effort-reward imbalance model of workplace stress with indicators of immune function. J Psychosom Res. 2016;91:1–8. doi: 10.1016/j.jpsychores.2016.10.003 [DOI] [PubMed] [Google Scholar]
  • 49.Violanti JM, Mnatsakanova A, Andrew ME, Allison P, Gu JK, Fekedulegn D. Effort–Reward Imbalance and Overcommitment at Work: Associations With Police Burnout. Police Q. 2018;21:440–60. doi: 10.1177/1098611118774764 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.von Helversen B, Wilke A, Johnson T, Schmid G, Klapp B. Performance benefits of depression: Sequential decision making in a healthy sample and a clinically depressed sample. J Abnorm Psychol. 2011;120:962–8. doi: 10.1037/a0023238 [DOI] [PubMed] [Google Scholar]
  • 51.Le Heron C, Plant O, Manohar S, Ang Y-S, Jackson M, Lennox G, et al. Distinct effects of apathy and dopamine on effort-based decision-making in Parkinson’s disease. Brain. 2018;141:1455–69. doi: 10.1093/brain/awy110 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52.Bonnelle V, Veromann K-R, Burnett Heyes S, Lo Sterzo E, Manohar S, Husain M. Characterization of reward and effort mechanisms in apathy. Journal of Physiology-Paris. 2015;109:16–26. doi: 10.1016/j.jphysparis.2014.04.002 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53.Cléry-Melin M-L, Schmidt L, Lafargue G, Baup N, Fossati P, Pessiglione M. Why Don’t You Try Harder? An Investigation of Effort Production in Major Depression. Laks J, editor. PLoS ONE. 2011;6:e23178. doi: 10.1371/journal.pone.0023178 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54.Foa EB, Huppert JD, Leiberg S, Langner R, Kichic R, Hajcak G, et al. The Obsessive-Compulsive Inventory: Development and validation of a short version. Psychol Assess. 2002;14:485–96. doi: 10.1037/1040-3590.14.4.485 [DOI] [PubMed] [Google Scholar]
  • 55.Mason O, Linney Y, Claridge G. Short scales for measuring schizotypy. Schizophr Res. 2005;78:293–6. doi: 10.1016/j.schres.2005.06.020 [DOI] [PubMed] [Google Scholar]
  • 56.Benjamin DJ, Berger JO, Johannesson M, Nosek BA, Wagenmakers E-J, Berk R, et al. Redefine statistical significance. Nature Human. Behaviour. 2018;2:6–10. doi: 10.1038/s41562-017-0189-z [DOI] [PubMed] [Google Scholar]
  • 57.Hauser TU, Moutoussis M, Dayan P, Dolan RJ. Increased decision thresholds trigger extended information gathering across the compulsivity spectrum. Transl Psychiatry. 2017;7:1–10. doi: 10.1038/s41398-017-0040-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58.Hauser TU, Moutoussis M, Iannaccone R, Brem S, Walitza S, Drechsler R, et al. Increased decision thresholds enhance information gathering performance in juvenile Obsessive-Compulsive Disorder (OCD). PLoS Comput Biol. 2017;13:e1005440. doi: 10.1371/journal.pcbi.1005440 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 59.Gillan CM, Kosinski M, Whelan R, Phelps EA, Daw ND. Characterizing a psychiatric symptom dimension related to deficits in goal-directed control. elife. 2016;5. doi: 10.7554/eLife.11305 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 60.Gillan CM, Morein-Zamir S, Urcelay GP, Sule A, Voon V, Apergis-Schoute AM, et al. Enhanced Avoidance Habits in Obsessive-Compulsive Disorder. Biol Psychiatry. 2014;75:631–8. doi: 10.1016/j.biopsych.2013.02.002 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 61.Vaghi MM, Cardinal RN, Apergis-Schoute AM, Fineberg NA, Sule A, Robbins TW. Action-Outcome Knowledge Dissociates From Behavior in Obsessive-Compulsive Disorder Following Contingency Degradation. Biol Psychiatry Cogn Neurosci Neuroimaging. 2019;4:200–9. doi: 10.1016/j.bpsc.2018.09.014 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 62.Vaghi MM, Luyckx F, Sule A, Fineberg NA, Robbins TW, De Martino B. Compulsivity Reveals a Novel Dissociation between Action and Confidence. Neuron. 2017;96:348–354.e4. doi: 10.1016/j.neuron.2017.09.006 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 63.Hauser TU, Allen M, Rees G, Dolan RJ. Metacognitive impairments extend perceptual decision making weaknesses in compulsivity. Sci Rep. 2017;7:1–10. doi: 10.1038/s41598-016-0028-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 64.Rouault M, Seow T, Gillan CM, Fleming SM. Psychiatric symptom dimensions are associated with dissociable shifts in metacognition but not task performance. Biol Psychiatry. 2018;84:443–51. doi: 10.1016/j.biopsych.2017.12.017 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 65.Bach DR, Moutoussis M, Bowler A, Dolan RJ. Predictors of risky foraging behaviour in healthy young people. Nat Hum Behav. 2020;4:832–43. doi: 10.1038/s41562-020-0867-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 66.Mata R, Wilke A, Czienskowski U. Foraging across the life span: is there a reduction in exploration with aging? Front Neurosci. 2013;7:53. doi: 10.3389/fnins.2013.00053 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 67.Lange K, Kühn S, Filevich E. " Just Another Tool for Online Studies”(JATOS): An Easy Solution for Setup and Management of Web Servers Supporting Online Studies. PLoS ONE. 2015;10:e0130834. doi: 10.1371/journal.pone.0130834 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 68.Ang Y-S, Lockwood P, Apps MAJ, Muhammed K, Husain M. Distinct Subtypes of Apathy Revealed by the Apathy Motivation Index. Gulinello M, editor PLoS ONE. 2017;12:e0169938. doi: 10.1371/journal.pone.0169938 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 69.Beck AT, Steer RA, Brown GK. Beck depression inventory-II San Antonio.1996;78:490–8. [Google Scholar]
  • 70.Spielberger CD. Manual for the State-Trait Anxiety Inventory STAI (form Y)(“self-evaluation questionnaire”). 1983. [Google Scholar]
  • 71.Liebowitz MR, Gorman JM, Fyer AJ, Klein DF (1985). Social phobia: Review of a neglected anxiety disorder. Archives of general psychiatry, 42(7), 729–736. [DOI] [PubMed] [Google Scholar]
  • 72.Michielsen HJ, De Vries J, Van Heck GL. Psychometric qualities of a brief self-rated fatigue measure. J Psychosom Res. 2003;54:345–52. doi: 10.1016/s0022-3999(02)00392-6 [DOI] [PubMed] [Google Scholar]
  • 73.Carpenter B, Gelman A, Hoffman MD, Lee D, Goodrich B, Betancourt M, et al. Stan: A Probabilistic Programming Language. J Stat Softw. 2017;76. doi: 10.18637/jss.v076.i01 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 74.Gabry J. Shinystan: interactive visual and numerical diagnostics and posterior analysis for Bayesian models. R package version 2015;2. [Google Scholar]
  • 75.Bürkner P-C. Advanced Bayesian multilevel modeling with the R package brms. arXiv preprint arXiv:170511123. 2017. [Google Scholar]
  • 76.Bürkner P-C, Charpentier E. Modeling Monotonic Effects of Ordinal Predictors in Regression Models. PsyArXiv. 2018. [cited 29 Jul 2019]. doi: 10.31234/osf.io/9qkhj [DOI] [PubMed] [Google Scholar]
  • 77.Bürkner P-C. brms: An R package for Bayesian multilevel models using Stan. J Stat Softw. 2017;80:1–28. [Google Scholar]
  • 78.Bürkner P-C, Vuorre M. Ordinal Regression Models in Psychology: A Tutorial. Adv Methods Pract Psychol Sci. 2019;2:77–101. doi: 10.1177/2515245918823199 [DOI] [Google Scholar]
  • 79.Revelle W. Procedures for personality and psychological research. Northwestern University, Evanston, IL. 2015. [Google Scholar]
  • 80.Hayes AF. Introduction to mediation, moderation, and conditional process analysis: A regression-based approach. Guilford Publications. 2017. [Google Scholar]
  • 81.Rosseel Y. lavaan: An R Package for Structural Equation Modeling. J Stat Softw. 2012;48:1–36. doi: 10.18637/jss.v048.i02 [DOI] [Google Scholar]
  • 82.Wickham H. ggplot2: elegant graphics for data analysis. Springer; 2016. [Google Scholar]
  • 83.Kassambara A. ggpubr:‘ggplot2’Based Publication Ready Plots (2018). R package version:02. [Google Scholar]
  • 84.Wickham H. R package version. 2017;2017 (1). [Google Scholar]
  • 85.Lüdecke D, Bartel A, Schwemmer C. Package ‘sjPlot.’ 2019. [Google Scholar]
  • 86.Scholl J, Kolling N, Nelissen N, Wittmann MK, Harmer CJ, Rushworth MF. The good, the bad, and the irrelevant: neural mechanisms of learning real and hypothetical rewards and effort. J Neurosci. 2015;35:11233–51. doi: 10.1523/JNEUROSCI.0396-15.2015 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 87.Gelman A, Carlin JB, Stern HS, Dunson DB, Vehtari A, Rubin DB. Bayesian data analysis Chapman and Hall/CRC. 2013. [Google Scholar]
  • 88.Gorsuch RL. Exploratory factor analysis: Its role in item analysis. J Pers Assess. 1997;68:532–60. doi: 10.1207/s15327752jpa6803_5 [DOI] [PubMed] [Google Scholar]
  • 89.Yong AG, Pearce S. A beginner’s guide to factor analysis: Focusing on exploratory factor analysis. Tutor Quant Methods Psychol. 2013;9:79–94. [Google Scholar]
  • 90.Hayton JC, Allen DG, Scarpello V. Factor Retention Decisions in Exploratory Factor Analysis: a Tutorial on Parallel Analysis. Organ Res Methods. 2004;7:191–205. doi: 10.1177/1094428104263675 [DOI] [Google Scholar]
  • 91.Moon K-W. Implementation of the “PROCESS” Macro [R package processR version 0.2.6]. Comprehensive R Archive Network (CRAN); 7 Jan 2021 [cited 20 May 2021]. Available from: https://CRAN.R-project.org/package=processR

Decision Letter 0

Roland G Roberts

11 Nov 2019

Dear Dr Scholl,

Thank you for submitting your manuscript entitled "Should I stick with it or move on? The effect of apathy and compulsivity on planning and stopping in sequential decision making" for consideration as a Registered Report by PLOS Biology.

Your manuscript has now been evaluated by the PLOS Biology editorial staff, as well as by an academic editor with relevant expertise, and I'm writing to let you know that we would like to send your submission out for external peer review.

IMPORTANT: The Academic Editor suggests that you include two additional items in the main manuscript before we sent it out to review. Please could you do this in addition to uploading the additional metadata:

AE's comments: "I would recommend that the authors make some minor revisions before in-depth review. First, the power analysis section that currently resides in the supplement should be moved into the main text. Second, I would recommend the authors include a table toward the end of the Stage 1 manuscript that shows the link between each hypothesis or sub-hypothesis, its corresponding Bayesian sampling plan (linked to that specific hypothesis and analysis), the specific analysis that will test that hypothesis, and contingent interpretation depending on different outcomes (making clear which outcomes would confirm or disconfirm the hypothesis). At present these links are not sufficiently clear for me, and would likely be an area of concern for reviewers."

ALSO IMPORTANT: Before we can send your manuscript to reviewers, we need you to complete your submission by providing the metadata that is required for full assessment. To this end, please login to Editorial Manager where you will find the paper in the 'Submissions Needing Revisions' folder on your homepage. Please click 'Revise Submission' from the Action Links and complete all additional questions in the submission questionnaire.

Please re-submit your manuscript within two working days, i.e. by Nov 13 2019 11:59PM. Do let us know if you need more time.

Login to Editorial Manager here: https://www.editorialmanager.com/pbiology

During resubmission, you will be invited to opt-in to posting your pre-review manuscript as a bioRxiv preprint. Visit http://journals.plos.org/plosbiology/s/preprints for full details. If you consent to posting your current manuscript as a preprint, please upload a single Preprint PDF when you re-submit.

Once your full submission is complete, your paper will undergo a series of checks in preparation for peer review. Once your manuscript has passed all checks it will be sent out for review.

Feel free to email us at plosbiology@plos.org if you have any queries relating to your submission.

Kind regards,

Roli Roberts

Roland G Roberts, PhD,

Senior Editor

PLOS Biology

Decision Letter 1

Roland G Roberts

7 Jan 2020

Dear Jacquie,

Thank you very much for submitting your manuscript "Should I stick with it or move on? The effect of apathy and compulsivity on planning and stopping in sequential decision making" for consideration as a Preregistered Research Article (Stage 1) at PLOS Biology. Your manuscript has been evaluated by the PLOS Biology editors, an Academic Editor with relevant expertise, and by three independent reviewers.

You'll see that the reviewers are broadly positive about your study, but each raises a number of requests for clarification, additional methodological detail, etc. In light of the reviews (below), we will not be able to accept the current version of the manuscript, but we would welcome re-submission of a much-revised version that takes into account the reviewers' comments. We cannot make any decision about publication until we have seen the revised manuscript and your response to the reviewers' comments. Your revised manuscript is also likely to be sent for further evaluation by the reviewers.

We expect to receive your revised manuscript within 2 months. Please note that I've also included at the foot of this email an updated version of the Guidelines, and the Stage 1 Template (which I don't think I sent you before). Many thanks for bearing with us as we iron out the wrinkles in our process for this new article type.

Please email us (plosbiology@plos.org) if you have any questions or concerns, or would like to request an extension. At this stage, your manuscript remains formally under active consideration at our journal; please notify us by email if you do not intend to submit a revision so that we may end consideration of the manuscript at PLOS Biology.

**IMPORTANT - SUBMITTING YOUR REVISION**

Your revisions should address the specific points made by each reviewer. Please submit the following files along with your revised manuscript:

1. A 'Response to Reviewers' file - this should detail your responses to the editorial requests, present a point-by-point response to all of the reviewers' comments, and indicate the changes made to the manuscript.

*NOTE: In your point by point response to to the reviewers, please provide the full context of each review. Do not selectively quote paragraphs or sentences to reply to. The entire set of reviewer comments should be present in full and each specific point should be responded to individually, point by point.

You should also cite any additional relevant literature that has been published since the original submission and mention any additional citations in your response.

2. In addition to a clean copy of the manuscript, please also upload a 'track-changes' version of your manuscript that specifies the edits made. This should be uploaded as a "Related" file type.

*Re-submission Checklist*

When you are ready to resubmit your revised manuscript, please refer to this re-submission checklist: https://plos.io/Biology_Checklist

To submit a revised version of your manuscript, please go to https://www.editorialmanager.com/pbiology/ and log in as an Author. Click the link labelled 'Submissions Needing Revision' where you will find your submission record.

Please make sure to read the following important policies and guidelines while preparing your revision:

*Published Peer Review*

Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out. Please see here for more details:

https://blogs.plos.org/plos/2019/05/plos-journals-now-open-for-published-peer-review/

*PLOS Data Policy*

Please note that as a condition of publication PLOS' data policy (http://journals.plos.org/plosbiology/s/data-availability) requires that you make available all data used to draw the conclusions arrived at in your manuscript. If you have not already done so, you must include any data used in your manuscript either in appropriate repositories, within the body of the manuscript, or as supporting information (N.B. this includes any numerical values that were used to generate graphs, histograms etc.). For an example see here: http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5

*Blot and Gel Data Policy*

We require the original, uncropped and minimally adjusted images supporting all blot and gel results reported in an article's figures or Supporting Information files. We will require these files before a manuscript can be accepted so please prepare them now, if you have not already uploaded them. Please carefully read our guidelines for how to prepare and upload this data: https://journals.plos.org/plosbiology/s/figures#loc-blot-and-gel-reporting-requirements

*Protocols deposition*

To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosbiology/s/submission-guidelines#loc-materials-and-methods

Thank you again for your submission to our journal. We hope that our editorial process has been constructive thus far, and we welcome your feedback at any time. Please don't hesitate to contact us if you have any questions or comments.

Best wishes,

Roli

Roland G Roberts, PhD,

Senior Editor

PLOS Biology

*****************************************************

REVIEWERS' COMMENTS:

Reviewer #1:

My understanding is that this is a "stage 1" paper that essentially provides context, study design, and pilot data. I review on that basis.

The authors plan to examine decision-making and compulsivity. The topic is very important and interesting. I enjoyed reading the paper and feel it has a number of novel aspects.

Comments:

1. The authors have already conducted a large study (n=449). Please clarify how the planned study is different - if it is simply collecting more data using an identical protocol, and the same recruitment pool (albeit in a new sample), then the presented data here aren't pilot data but a large part of an already substantially advanced study. This is fine but if so I don't think this is a 'stage 1' paper in the sense I have understood it (my understanding is that stage 1 is essentially "pre-registration" of methods for a study you plan to do in advance of doing the study). Put differently, if the study is already quite advanced in progress, shouldn't all the data just be reported in one paper, as it isn't "pre-registration"?

2. "We collected questionnaires with measures covering a wide range of psychiatric symptoms....."

Only one type of compulsivity is considered in the study design (OCD) as far as I can discern. I appreciate some papers in the literature uses OCD symptom scales (e.g. OCI; Padua...) as a putative 'trans-diagnostic' measure but it isn't really, as all the items are specifically to do with OCD and not the vast range of other types of compulsive symptoms (including those mentioned by the authors). I would encourage the authors to incorporate a broader range of symptoms into their 'compulsivity' measure. Or if they focus on OCD, this is fine, but actually it should be referred to as "OCD-related compulsivity", rather than a trans-diagnostic measure.

3. More detail should be provided about how the participants (and future participants) were recruited and selected. People doing online platforms may be unrepresentative of the general population. Also give thought to: are data collected online clinically valid? Is the cognitive task valid administered over the Internet? This may be the case but please provide more / clearer information. It would be important to list clearly as a limitation that online data collection is not the gold standard and may not generalise to findings collected in-person using trained raters, objective measures, and supervised cognitive tests.

4. Consider whether exclusion based on questionnaire responses, as listed, might be too strict. You will probably knockout people in your sample who are particularly compulsive and/or have attentional abnormalities (common in the setting of compulsivity). Best would be to undertake analyses with full datasets and compare findings using these more rigorous criteria. That way you address the issue but don't have to worry about biasing your sample.

5. Statistical approach is very convoluted for the decision-making task and its outcome measures. What will these measures mean for clinicians and patients? How easy will it be for other groups to replicate any results, unless they have a PhD in statistics? How many complex statistical assumptions are made that may or may not be valid? This is the reviewer's clarion call to the authors for well-validated straight forward decision-making task with clear easy-to-understand outcome measures. The reviewer appreciates he appears to be fighting a losing battle in the field of decision-making in this regard. If it really is necessary to use such a complex approach, please explain why, and provide an overview of the steps in language that a typical person reading a psychiatric or neuroscience journal would understand; aside from also providing the very detailed information too.

6. CFA will be done in independent data-set; correct?

7. Have all confounds be considered - what about IQ, years of education, psychoactive medications, etc.?

Reviewer #2:

[identifies himself as Jiaxiang Zhang]

[The importance of the research question(s)]

The current pre-reg report aims to collect behavioural responses from a large online sample (>650). The sequential decision-making task to be used is from the authors' previous study, which ensures feasibility of the proposed study and validity of data collection. The large online sample enables addressing new important questions, specifically how apathy and compulsivity traits may relate to differences in choice behaviour across participants.

[The logic, rationale, and plausibility of the proposed hypotheses]

The ms provided a good rationale for the proposed research, including the need for using transdiagnostic trait measures. Three main hypotheses have been identified, relating (1) apathy to decision inertia, (2) compulsivity to cost insensitivity and (3) compulsivity to self-report measures.

I hope the authors could clarify the following two issues:

1. The ms referred to clinical populations (OCD and depression) when justifying the hypotheses. Can the current and proposed results rule out an alternative explanation that the trait-performance linear association is mainly/disproportionally contributed by those individuals with more extreme trait measures (as in Fig. 2A)?

2. The link between high apathy and excessive effort is indeed counterintuitive. I wonder why, in the current sequential decision task, participants with high apathy trait did not exhibit any avolitional types of behaviour, e.g., longer RT in the first decision (initial search) of each trial?

[ The soundness and feasibility of the methodology and analysis pipeline ]

The authors conducted rigorous analysis on an existing online dataset, with detailed description of methods and power analysis for the proposed study. The protocol (paradigm, hierarchical modelling and analyses) is technically sound and appropriate for testing the stated hypotheses.

The author should provide a more explicit list on all the differences in data/analysis between the discovery (existing) and confirmation samples, as they scattered across the current ms (one-tailed/two-tailed, changes in questionnaires, changes in schedules, etc.).

[ Whether the clarity and degree of methodological detail is sufficient to exactly replicate the proposed experimental procedures and analysis pipeline.]

I am confused why "… 10 random task schedules …" were constructed. Will a subset of participant be tested with an identical task sequence, or will each participants be presented with an unique schedule?

[ Whether the authors have pre-specified sufficient outcome-neutral tests for ensuring that the results obtained are able to test the stated hypotheses, including positive controls and quality checks]

Yes

[Minor]

P25. "Participants were did not adapt …"

Reviewer #3:

[identifies himself as Christoph W. Korn]

Scholl et al. propose an exciting study for a registered report: In a large online study, they use a validated sequential risky decision-making task in conjunction with a computational model to dissociate effects of apathy and compulsions and then link these to individual differences based on a transdiagnostic clustering of various questionnaires (pilot sample n>440; proposed sample n=650). I and a PhD student who helped me compile this review are generally in favor of seeing this study conducted. Here is a short assessment of the criteria for Stage 1 enlisted in the reviewer instructions (followed by more detailed specific points):

a) The research questions are important from a theoretical and an applied view (e.g., theoretical: how well can constructs such as apathy and compulsivity as derived from questionnaires be linked to features from computational analyses of decision-making tasks?; applied: how could this inform research on psychiatric spectrum conditions?).

b) The logic of the hypotheses is rather valid and clear. The authors could make it a bit easier for the readers to understand the modelling approach (see comments below).

c) The study is definitively sound and feasible. The current data and results are already quite substantial and the overall analyses are appropriate and sophisticated.

d) The methodological descriptions are detailed and allow for an exact replication.

e) The authors have conducted quality tests for their pilot data and one can assume that they will do the same for the main study. Some more quality checks are proposed below.

Specific points:

1) Registered report:

a. Motivation for registered report: The authors do not clarify why they want to conduct a registered report (rather than publishing their current data as they are). Of course, there are numerous advantages to a registered report. We feel that the specific reasons should be stated briefly.

b. Approach for registered analyses: Overall the methods are quite clearly described for the current data. But is unclear to us if the authors want to change anything for the new data collection and analyses (except for some changes mentioned in the supplementary file, e.g., changing the number of questionnaires). Put differently, have we learned anything from the current data that would suggest potential improvements for the new study? If not, that's fine and it would help the readers to know what exactly changed between the first and second study, even nothing changed.

2) Computational model and hypotheses:

a. Description in introduction and main text: It seems to us that the authors want to spare readers the technical details of the model in the main text. This comes at the cost of a rather blurry understanding of how the hypotheses are tested. For example, the second paragraph of the introduction lists what the model can achieve but this remains rather fanciful without reading the methods, which makes the paper's methods appear less robust than they actually are. Some more technical details would help (similar to the very helpful Table 2). In particular, the authors should state clearly in the main text how they can dissociate H1 and H2 since both are superficially linked to "searching longer for a better option." Even readers not versed in modelling could understand that H1 relates more to the number of previous searches while H2 relates to the search costs, which vary across trials. Also, it should be made clear that H3 is based on asking participants a separate (meta-cognitive) question about the task.

b. Alternative models: Do the authors test any potentially simpler or more complex models (irrespective of the link to individual differences)? It seems to us that the authors should first show that a given model component is relevant for the whole group of participants before going into questions of individual differences.

c. Split of behavior in 1st decision to search and consecutive decisions: Could the authors give a better rationale for this in the main text. Are the RTs of the 1st decision meaningful given that there was a forced (and jittered) delay?

d. "Decision" inertia: We were initially a bit confused by this term since participants with "decision inertia" actually end up making more risky decisions. They rather seem to have a "commitment" or "closure" inertia because they don't want to commit to a particular outcome. A clarification early on would help.

e. Sunk-cost effects: Related to the previous point. The relation of task behavior to sunk-cost effects could be clearer early on.

f. Report of individual effects: When reading the results I found it initially confusing that the individual effects were reported as mean parameter estimates in a model rather than as correlations. Maybe correlations would be helpful because they allow more intuitive assessments of effect sizes.

g. "Model-free" descriptive statistics and posterior predictive checks: Some more "model-free" diagnostics would be helpful. For example, what is the average number of searches per trial? Also, posterior predictive checks for the main model variables would be needed - as well as scatter plots for the main hypotheses (and ideally individual data points or violin plots).

3) Analyses of questionnaires:

a. 4 factors: Why did the authors restrict themselves to 4 factors? How much variance do these explain?

b. Influence of "social anxiety:" Do the authors have explanations for the negative effect of social anxiety on RTs (in Figure 3B & 4B)?

4) Various minor comments:

a. The introduction is very short and mostly lists the task and the hypotheses.

b. It could be clearer why the authors use an individual differences approach. For example, another valid approach would be to test to groups of patients that differ in apathy and compulsivity (e.g., MDD v. OCD patients).

c. How long is the experiment? How do the authors make sure that the online participants don't interrupt the experiment? We're aware that Prolific includes a 'Maximum time allowed', but we assume that many people won't be aware of this and therefore might be concerned that participants take long breaks between parts of the experiment (e.g., take a long break between the behavioral task and the questionnaires)

d. The authors exclude persons with too little variance in their answers, which is completely fine. However, since so many questionnaires are available, it could be interesting to explore the questionnaire variables of these "excluded" participants. In other words: is there anything interesting/unusual about the people who had little variance?

e. In the power analyses presented in Figure 5, how can power decrease with larger sample size (e.g., 'Compulsivity x Chasing' has ~96% power at 450 participants, but only ~87% power at 550 participants)? Is this an artefact of the randomness in the simulations? If so, would more simulated runs be necessary to avoid this?

f. Although we appreciate the authors' attempt to make their research question especially relevant to current debates, it seems unlikely that sequential decision-making has become dramatically more common for animals and humans since 01.01.2000 (see 2nd sentence in introduction and their explicit emphasis on the 21st century). It should also be added that humans are animals, so that distinction seems somewhat redundant

g. The authors state multiple times that they want to get more ecological validity and that sequential decision-making is more realistic and that the dimensional approach is more realistic, but the last point is not so obvious: why exactly is the dimensional approach more realistic than just the questionnaires? Isn't the whole point of a well-designed questionnaire to learn something about someone that can help make predictions about their real-life behavior?

UPDATED GUIDELINES AND TEMPLATE:

Guidelines for Authors

Registered Reports/Pre-Registered Research Articles are a form of empirical article offered at PLOS Biology in which study rationale, methods and proposed analyses are reviewed prior to research being conducted. High quality protocols are reviewed for technical soundness of the proposed methodology, and provisionally accepted for publication before data collection commences. This format of article is designed to minimise publication and reporting bias, while also maximising study quality by focusing peer review on the importance of the research question/theory and rigour of the proposed methodology. It also allows complete flexibility to conduct exploratory (unregistered) analyses and report serendipitous findings. Pre-Registered Research Articles are offered across the full scope of empirical research at PLOS Biology.

Editorial and Peer Review Process for Pre-Registered Research Articles

Initial submissions (Stage 1 study proposals) will be triaged by the editorial team for importance of the research question using PLOS Biology’s general criteria for publication. Please see below for the template that should be followed for Stage 1 study proposals. Those that pass editorial triage will be invited for full submission (to complete additional manuscript details) and sent for in-depth peer review to further assess importance of the research question and to evaluate the technical soundness of the proposed study design and methodology (Stage 1).

Following Stage 1 peer review, manuscripts will be rejected, offered the opportunity to revise the study proposal, or accepted in principle. Study proposals that meet our high standards of importance and scientific rigour will be issued an in-principle acceptance decision (perhaps after revision/s, as needed), indicating that the article will be published pending completion of the study. Stage 1 protocols are not published following upon an in-principle acceptance. Instead they are held and integrated into a single completed ‘Pre-Registered Research Article’ following review and acceptance of the final Stage 2 manuscript.

Following a Stage 1 in-principle acceptance decision, authors are required to register their approved protocol on the Open Science Framework (https://osf.io/) or other recognised repository, either publicly or under private embargo until submission of the Stage 2 manuscript. Accepted protocols can be quickly and easily registered using a tailored mechanism for Registered Reports on the Open Science Framework: https://osf.io/rr/ Authors then proceed to conduct the study, adhering exactly to the peer-reviewed and approved Stage 1 study design. When the study is complete the authors will submit their finalised manuscript, including Results and Discussion sections, for review (Stage 2). Editorial decisions will not be based on the perceived importance or novelty of the results obtained when completing the Stage 2 study.

Any deviation from approved Stage 1 study procedures (after in-principle acceptance), regardless of how minor it may seem to the authors, could lead to rejection of the manuscript at Stage 2. In cases where the pre-approved protocol is altered after in-principle acceptance due to unforeseen circumstances (e.g. change of equipment or unanticipated technical error), the authors must consult the editors immediately for advice, and prior to the completion of data collection. Minor changes to the protocol may be permitted according to editorial discretion. In such cases, the deviation must be reported in the Stage 2 submission.

Submitting a Stage 1 Manuscript

Stage 1 submissions should include the manuscript (details below) and a cover letter.

Authors are welcome to submit presubmission inquires for advice on the likely suitability of a study for PLOS Biology by emailing biology_editors@plos.org. However, please note that the editorial team cannot officially commit to sending manuscripts for in-depth review until a complete Stage 1 initial submission has been submitted online.

The cover letter should include:

A brief scientific case for consideration.

A statement confirming that all necessary support (e.g. funding, facilities) and approvals (e.g. ethics) are in place for the proposed research. Note that manuscripts will generally be considered only for studies that are able to commence immediately; however authors with alternative plans are encouraged to contact the journal editors for advice.

An anticipated timeline for completing the study if the study proposal is approved and accepted.

A statement confirming that, following Stage 1 in principle acceptance, the authors agree to register their approved protocol on the Open Science Framework (https://osf.io/) or other recognised repository, either publicly or under private embargo until submission of the Stage 2 manuscript. Accepted protocols can be quickly and easily registered using a tailored mechanism for Registered Reports on the Open Science Framework: https://osf.io/rr/

A statement confirming that the authors agree to share their data, in accordance with the PLOS Data Availability Policy, and laboratory log, if needed, for all published results.

Stage 1 Manuscript Template (Study Proposal)

Stage 1 manuscripts should include the following sections:

Abstract

Brief note of question being tested, the relevance, and the proposed investigative approach. (Note: This section should have an additional outcome paragraph added in Stage 2 manuscripts).

Introduction

A review of the relevant literature that motivates the research question and a full description of the experimental aims and hypotheses. Please make sure to enumerate the specific hypotheses. Please note that this section cannot be altered at Stage 2 (see below) after a Stage 1 in-principle acceptance.

Materials and Methods

Please note that this section cannot be altered at Stage 2 (see below) after a Stage 1 in-principle acceptance.

Protocol details (repeat as necessary for all protocols being proposed): Sampling, Cell Lines / Organisms (i.e. experimental population), Materials & Reagents, Procedure/Intervention, Deliverables. Experimental procedures should be provided in sufficient detail to allow another researcher to repeat the methodology exactly, without requiring further information.

Sampling plan (e.g. power calculations or Bayesian sampling methods etc.) should be included unless clearly not appropriate. Please include details of criteria for data inclusion and exclusion (e.g. outlier extraction); procedures for objectively defining exclusion criteria due to technical errors or for any other reasons must be specified, including details of how and under what conditions data would be replaced. Please also detail when data collection would cease e.g. sample size, number of observations etc.

Proposed analysis pipeline, including all preprocessing steps, and a precise description of all planned analyses, including appropriate correction for multiple comparisons. Any covariates or regressors must be stated. Where analysis decisions are contingent on the outcome of prior analyses, these contingencies must be specified and adhered to. Only pre-planned analyses can be reported in the main Results section of Stage 2 manuscripts. However, unplanned exploratory analyses will be admissible in a separate section of the Results (see Stage 2 details below).

Statistics:

When relevant, studies involving Neyman-Pearson inference must include a statistical power analysis. Estimated effect sizes should be justified with reference to the existing literature. Since publication bias overinflates published estimates of effect size, power analysis must be based on the lowest available or meaningful estimate of the effect size. The a priori power must be 0.9 or higher for all proposed hypothesis tests. In the case of highly uncertain effect sizes, a variable sample size and interim data analysis will be permissible but with inspection points stated in advance, appropriate Type I error correction for ‘peeking’ employed, and a final stopping rule for data collection outlined.

Methods involving Bayesian hypothesis testing are particularly encouraged. For studies involving analyses with Bayes factors, the predictions of the theory must be specified so that a Bayes factor can be calculated. Authors should indicate what distribution will be used to represent the predictions of the theory and how its parameters will be specified. For example, will you use a uniform up to some specified maximum, or a normal/half-normal to represent a likely effect size, or a JZS/Cauchy with a specified scaling constant? For inference by Bayes factors, authors must be able to guarantee data collection until the Bayes factor is at least 10 times in favour of the experimental hypothesis over the null hypothesis (or vice versa). Authors with resource limitations are permitted to specify a maximum feasible sample size at which data collection must cease regardless of the Bayes factor, however to be eligible for advance acceptance this number must be sufficiently large that inconclusive results at this sample size would nevertheless be of major importance.

Full descriptions must be provided of any outcome-neutral criteria that must be met for successful testing of the stated hypotheses. Such quality checks might include the absence of floor or ceiling effects in data distributions, positive controls, or other quality checks.

Any description of prospective methods or analysis plans should be written in future tense. For the Stage 2 manuscript, once the study is complete, these instances of future tense should be changed to past tense.

Summary Table

Please include a summary table that aligns each research question with the hypothesis/es used to answer the question, the sampling plan for each hypothesis (e.g. power analysis, where applicable), the specific statistical analysis/es that will be used to test the hypothesis, and a pre-specification of which outcomes will confirm or disconfirm the hypothesis (to varying degrees of strength where multiple analyses with different possible outcomes are used to interrogate one hypothesis).

Timeline

Present anticipated timeline for completion of the study and proposed resubmission date if Stage 1 review is successful. Extensions to this deadline can be discussed with the editor.

Pilot Data

Optional. Can be included to establish proof of concept, effect size estimations, or feasibility of proposed methods and can include details of any preliminary data that have already been obtained (approach, materials and methods, results, analytical observations etc.). Any pilot studies will be published with the final version of the manuscript and will need to be clearly distinguished from data subsequently obtained at Stage 2 from the pre-approved experiments.

Secondary Analyses

Optional. If the study proposes secondary analyses of existing databases, please provide full details of the data to be analysed, its origin and any relevant information and citations with respect to the origin of the dataset and any previous analyses that have been performed. Please make clear what extent of prior observation you have had (secondary analysis of existing data may be bias-prone, which should be avoided). If this is a proposed replication study, full details of replication approach should be provided.

Data Availability Plan

Provide full details of where/how data and/or code produced will be shared (in line with the PLOS Data & Code sharing Policies).

Ethical Approval Plan

Provide details of ethical approval for animal and human subject research.

Decision Letter 2

Roland G Roberts

7 Jul 2020

Dear Dr Scholl,

Thank you very much for submitting your revised Stage 1 manuscript "Should I stick with it or move on? The effect of apathy and compulsivity on planning and stopping in sequential decision making" for consideration as a Preregistered Research Article at PLOS Biology. Your revisions have been assessed by the reviewers, and I've discussed their comments with the Academic Editor.

The reviews of your manuscript are appended below. I have discussed the concern raised by reviewer #1 with the Academic Editor, who is satisfied (as are the other reviewers) that your Stage 1 Protocol meets our criteria for importance of research question and technical soundness of the study proposal. We would therefore like to invite you to complete the study, as proposed, and submit the Stage 2 manuscript. Please carefully read all the following information.

*Explanation of Decision*

This is a Stage 1 'in-principle acceptance' decision, with a commitment to publish the final Stage 2 Preregistered Research Article (after revision, if needed), pending successful completion of the study according to these Stage 1 approved methods and analytic procedures, as well as an evidence-based interpretation of the results. Please see here for review criteria for Stage 2 manuscripts:

https://journals.plos.org/plosbiology/s/reviewer-guidelines#loc-reviewing-preregistered-research-articles

Editorial decisions will not subsequently be based on the perceived importance or novelty of the results obtained during the Stage 2 study. It is critical however that you adhere exactly to this approved Stage 1 study design when performing the study. Any deviation from these experimental procedures could lead to rejection of the manuscript at Stage 2. Please consult the editors immediately for advice if you need to alter this approved study plan.

**IMPORTANT**: Please follow the link below for important information regarding the Stage 2 manuscript template and review criteria. Please carefully read the guidelines on Stage 2 data collection BEFORE performing your study and completing your Stage 2 manuscript.

AUTHOR GUIDELINES: https://plos.io/AuthorGuidelines

*Depositing this Stage 1 Protocol*

PLOS Biology does not publish Stage 1 Protocols immediately following an in-principle acceptance. Instead they are held and integrated into a single, completed 'Preregistered Research Article' following review and acceptance of the final Stage 2 manuscript. You are however required to register this approved Stage 1 Protocol with the Center for Open Science (https://cos.io/prereg/) or another recognised repository. This may be done publicly or under private embargo until submission of the Stage 2 manuscript. Stage 1 Protocols can be quickly and easily registered using a tailored mechanism for Registered Reports (https://osf.io/rr/). Please do this now. You will need to include the URL to this deposited protocol in your Stage 2 manuscript.

*Timeline*

We understand that carrying out the study will require a significant length of time and and are willing to allow you up to nine months to perform the study (we note that you stipulate six months in your manuscript, but we assume that the ongoing pandemic may result in some delays). Please email us at 'plosbiology@plos.org' to discuss this if you have any questions or concerns, or to discuss an alternate timeline.

At this stage, your manuscript remains formally under active consideration at our journal. Please notify us by email if you do not wish to submit a Stage 2 manuscript or wish to pursue publication elsewhere, so that we may end consideration of the manuscript at PLOS Biology.

*Resubmission Checklist*

Before submitting the Stage 2 manuscript, please review the following resubmission checklist: https://plos.io/Biology_Checklist

To submit a revised version of your manuscript, please go to https://www.editorialmanager.com/pbiology/ and log in as an Author. Click the link labelled 'Submissions Needing Revision' where you will find your submission record.

Please make sure to read the following important policies and guidelines while preparing your revision:

*Published Peer Review*

Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out. Please see here for more details:

https://blogs.plos.org/plos/2019/05/plos-journals-now-open-for-published-peer-review/

*PLOS Data Policy*

Please note that as a condition of publication, PLOS' data policy (http://journals.plos.org/plosbiology/s/data-availability) requires that you make available all data used to draw the conclusions arrived at in your manuscript. Please note that for this article type, the raw data itself should be archived and made freely available in a public repository rather than submitted as supplementary material. Please make sure to read the Stage 2 submission guidelines online regarding how this data should be annotated and appropriately time stamped to show that data was collected after this Stage 1 in-principle acceptance and not before.

*Blot and Gel Data Policy*

We require the original, uncropped and minimally adjusted images supporting all blot and gel results reported in an article's figures or Supporting Information files. We will require these files before a manuscript can be accepted so please prepare them now, if you have not already uploaded them. Please carefully read our guidelines for how to prepare and upload this data: https://journals.plos.org/plosbiology/s/figures#loc-blot-and-gel-reporting-requirements

*Protocols deposition*

To enhance the reproducibility of your results, we recommend that, if applicable, you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosbiology/s/submission-guidelines#loc-materials-and-methods

Thank you again for your submission to PLOS Biology. We hope that our editorial process has been constructive thus far, and we welcome your feedback at any time. Please don't hesitate to contact us if you have any questions or comments.

Sincerely,

Roli Roberts

Roland G Roberts, PhD,

Senior Editor

PLOS Biology

*****************************************************

REVIEWERS' COMMENTS:

Reviewer #1:

My only remaining point is (also raised in my original review): is this really a pre-registration?

The rationale for a pilot in several hundred people in order to do a power calculation to collect more data is not convincing - it is pretty obvious that a study in 700 or so people would have sufficient power to test specific hypotheses - e.g. based on power calculations not requiring data (e.g. assuming a medium or small effect size). The idea of collecting two samples to cross-validate results is of course excellent and to be applauded. But isn't this really a single study that should be submitted as a single full data paper? Pre-registration is done before a research study, not after it. As such, publishing this as a pre-registration paper seems inappropriate/inaccurate to me. But I leave that judgement to the editor.

The authors have addressed the other issues raised in my original review. For this reason, my recommendation is accept, but subject to the editor confirming that this is a pre-registration paper; and if not, I recommend it be published as a standard paper.

For future work, I would encourage the authors to think beyond only OCD in terms of compulsivity. Are the factor scores provided by e.g. Padua/OCI (e.g. washing/checking, contamination fears, ordering, thoughts of harm to self/others) really telling us about trans-diagnostic compulsivity? Many of these aren't relevant to any other compulsive disorder apart from OCD.

Reviewer #2:

[identifies himself as Jiaxiang Zhang]

The authors addressed all my concerns in this revision. I look forward to the results of this pre-reg study.

Reviewer #3:

[identifies himself as Christoph W. Korn]

We would like to thank the authors for addressing all of our comments. For example, Figure R5 is very helpful. The responses regarding alternative models and the transdiagnostic approach are convincing and very well taken.

We wish the authors all the best for the data collection and look forward to seeing the results.

Decision Letter 3

Roland G Roberts

21 Jul 2021

Dear Jacquie,

Many thanks for submitting the Stage 2 version of your manuscript "Should I stick with it or move on? The effect of apathy and compulsivity on planning and stopping in sequential decision making" for consideration as a Preregistered Research Article at PLOS Biology.

IMPORTANT: In order to make the path as smooth as possible, the Academic Editor has asked me to get you to do the following things before we send it back out for Stage 2 review:

1. Please include the OSF URL for the registered Stage 1 MS, instead of your current placeholder. If the OSF deposition is under embargo, you now need to make it public, especially as the reviewers may need to cross-check the current version against it. (see https://plos-marketing.s3.amazonaws.com/Marketing/Biology+Preregistered+Articles+Guidelines+for+Authors.pdf). I will also include this link in the reviewer invitations.

2. Note that I will then check that the OSF-registered version only differs from our Stage 1 accepted version with regard to the approved change (a tweak to exclusion criteria) that you made after acceptance.

3. Please provide a new "Track Changes" (mark-up version) that also includes tracking of all text changes in the Abstract, Introduction and Method sections (at the moment you've done so for the Methods section, but there are also changes to the Abstract that aren't tracked; I haven't checked the Introduction). Ideally you should mark up the actual changes, not just highlight the changed paragraphs. Again, I will re-check this before review.

4. The Academic Editor suggests that you update Table 4 (the design table) to include an extra column to the right called "Outcome" which simply reports whether the hypothesis was confirmed or not confirmed. The reviewers and readers will probably find this helpful.

We expect to receive your revised manuscript within 1 week.

Please email us (plosbiology@plos.org) if you have any questions or concerns, or would like to request an extension. At this stage, your manuscript remains formally under active consideration at our journal; please notify us by email if you do not intend to submit a revision so that we may end consideration of the manuscript at PLOS Biology.

**IMPORTANT - SUBMITTING YOUR REVISION**

In addition to a clean copy of the manuscript, please also upload a 'track-changes' version of your manuscript that specifies the edits made. This should be uploaded as a "Related" file type.

*Resubmission Checklist*

When you are ready to resubmit your revised manuscript, please refer to this resubmission checklist: https://plos.io/Biology_Checklist

To submit a revised version of your manuscript, please go to https://www.editorialmanager.com/pbiology/ and log in as an Author. Click the link labelled 'Submissions Needing Revision' where you will find your submission record.

Please make sure to read the following important policies and guidelines while preparing your revision:

*Published Peer Review*

Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out. Please see here for more details:

https://blogs.plos.org/plos/2019/05/plos-journals-now-open-for-published-peer-review/

*PLOS Data Policy*

Please note that as a condition of publication PLOS' data policy (http://journals.plos.org/plosbiology/s/data-availability) requires that you make available all data used to draw the conclusions arrived at in your manuscript. If you have not already done so, you must include any data used in your manuscript either in appropriate repositories, within the body of the manuscript, or as supporting information (N.B. this includes any numerical values that were used to generate graphs, histograms etc.). For an example see here: http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5

*Blot and Gel Data Policy*

We require the original, uncropped and minimally adjusted images supporting all blot and gel results reported in an article's figures or Supporting Information files. We will require these files before a manuscript can be accepted so please prepare them now, if you have not already uploaded them. Please carefully read our guidelines for how to prepare and upload this data: https://journals.plos.org/plosbiology/s/figures#loc-blot-and-gel-reporting-requirements

*Protocols deposition*

To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols

Thank you again for your submission to our journal. We hope that our editorial process has been constructive thus far, and we welcome your feedback at any time. Please don't hesitate to contact us if you have any questions or comments.

Best wishes,

Roli

Roland Roberts

Senior Editor

PLOS Biology

rroberts@plos.org

Decision Letter 4

Roland G Roberts

25 Oct 2021

Dear Jacquie,

Thank you for submitting your revised Preregistered Research Article entitled "Should I stick with it or move on? The effect of apathy and compulsivity on planning and stopping in sequential decision making" for publication in PLOS Biology. I have now obtained advice from two of the original reviewers and have discussed their comments with the Academic Editor. 

Based on the reviews, we will probably accept this manuscript for publication, provided you satisfactorily address the remaining points raised by the reviewers. Please also make sure to address the following data and other policy-related requests.

IMPORTANT: Please attend to the following:

a) Please attend to the reviewers' remaining requests. Note that the Academic Editor asked me to transmit the following important caveat: "Concerning the comment from rev #2: 'Although exploratory analyses are clearly marked as they are, I wonder if the authors could unpack the results in Intro/Discussion a bit more, as they can be as useful as confirmatory results for future studies.' This is a useful suggestion but it's very important to make clear to the authors (in the decision letter) to interpret this remark solely in terms of the Discussion, NOT the Intro. We don't want the authors making any further changes to the Introduction at this point unless it's to correct factual errors or typos. So this kind of contextualisation and interpretation should be reserved for the Discussion."

b) Please could you remove the first sentence of your title to make it more succinct and direct: "The effect of apathy and compulsivity on planning and stopping in sequential decision making"

c) Please address my Data Policy requests below; specifically, please supply numerical values underlying Figs 1DEF, 2ABCD, 3ABC, 4, 5ABC, 6, S1ABCD, S2ABCDEF, S3ABC, S4ABCD, S5, and cite the location of the data clearly in each relevant Fig legend ("The data underlying this Figure may be found at https://..."). You currently say that these data will be made available on OSF upon acceptance, but this is not satisfactory; we cannot editorially accept your paper until we have checked data availability. You will also note that both reviewers raise this point.

As you address these items, please take this last chance to review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the cover letter that accompanies your revised manuscript.

We expect to receive your revised manuscript within two weeks.

To submit your revision, please go to https://www.editorialmanager.com/pbiology/ and log in as an Author. Click the link labelled 'Submissions Needing Revision' to find your submission record. Your revised submission must include the following:

-  a cover letter that should detail your responses to any editorial requests, if applicable, and whether changes have been made to the reference list

-  a Response to Reviewers file that provides a detailed response to the reviewers' comments (if applicable)

-  a track-changes file indicating any changes that you have made to the manuscript. 

NOTE: If Supporting Information files are included with your article, note that these are not copyedited and will be published as they are submitted. Please ensure that these files are legible and of high quality (at least 300 dpi) in an easily accessible file format. For this reason, please be aware that any references listed in an SI file will not be indexed. For more information, see our Supporting Information guidelines:

https://journals.plos.org/plosbiology/s/supporting-information  

*Published Peer Review History*

Please note that you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out. Please see here for more details:

https://blogs.plos.org/plos/2019/05/plos-journals-now-open-for-published-peer-review/

*Early Version*

Please note that an uncorrected proof of your manuscript will be published online ahead of the final version, unless you opted out when submitting your manuscript. If, for any reason, you do not want an earlier version of your manuscript published online, uncheck the box. Should you, your institution's press office or the journal office choose to press release your paper, you will automatically be opted out of early publication. We ask that you notify us as soon as possible if you or your institution is planning to press release the article.

*Protocols deposition*

To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols

Please do not hesitate to contact me should you have any questions.

Best wishes,

Roli

Roland G Roberts, PhD,

Senior Editor,

rroberts@plos.org,

PLOS Biology

------------------------------------------------------------------------

DATA POLICY:

You may be aware of the PLOS Data Policy, which requires that all data be made available without restriction: http://journals.plos.org/plosbiology/s/data-availability. For more information, please also see this editorial: http://dx.doi.org/10.1371/journal.pbio.1001797 

Note that we do not require all raw data. Rather, we ask that all individual quantitative observations that underlie the data summarized in the figures and results of your paper be made available in one of the following forms:

1) Supplementary files (e.g., excel). Please ensure that all data files are uploaded as 'Supporting Information' and are invariably referred to (in the manuscript, figure legends, and the Description field when uploading your files) using the following format verbatim: S1 Data, S2 Data, etc. Multiple panels of a single or even several figures can be included as multiple sheets in one excel file that is saved using exactly the following convention: S1_Data.xlsx (using an underscore).

2) Deposition in a publicly available repository. Please also provide the accession code or a reviewer link so that we may view your data before publication. 

Regardless of the method selected, please ensure that you provide the individual numerical values that underlie the summary data displayed in the following figure panels as they are essential for readers to assess your analysis and to reproduce it: 1DEF, 2ABCD, 3ABC, 4, 5ABC, 6, S1ABCD, S2ABCDEF, S3ABC, S4ABCD, S5. NOTE: the numerical data provided should include all replicates AND the way in which the plotted mean and errors were derived (it should not present only the mean/average values).

IMPORTANT: Please also ensure that figure legends in your manuscript include information on where the underlying data can be found, and ensure your supplemental data file/s has a legend.

Please ensure that your Data Statement in the submission system accurately describes where your data can be found.

------------------------------------------------------------------------

DATA NOT SHOWN?

- Please note that per journal policy, we do not allow the mention of "data not shown", "personal communication", "manuscript in preparation" or other references to data that is not publicly available or contained within this manuscript. Please either remove mention of these data or provide figures presenting the results and the data underlying the figure(s).

------------------------------------------------------------------------

REVIEWERS' COMMENTS:

Reviewer #2:

[identifies himself as Jiaxiang Zhang]

I have reviewed the Stage 1 report of this study. It is great to see the confirmatory results in the current report. In this manuscript, the intro, rationale, and hypotheses are the same as the stage 1 protocol, making it a high-quality stage 2 submission. The authors did a great job in adhering to original experimental procedures and analyses plans. In the case of potential inconsistency (e.g., factor analyses), the authors provided convincing evidence to stick with the original analysis plan. Overall, beyond the obvious interest to the decision-making/cognitive psychology community, I believe this study will set a nice example on how to combine a novel experimental paradigm with rigorous pre-registered analyses. Below are just a few minor comments.

1. Table 3. Although exploratory analyses are clearly marked as they are, I wonder if the authors could unpack the results in Intro/Discussion a bit more, as they can be as useful as confirmatory results for future studies. What do they mean, what can we learn from them, or how can they be extended in future studies? For example, cost is not a reliable predictor (which I think is the same predictor in hypothesis 2A?) of compulsivity, but "chasing" is. There are however other task-based measures that, together, yield a better predictive model for compulsivity. Does this mean trait measure of compulsivity is associated with multifaceted behaviour and meta-cognitive processes, instead of specific correlations highlighted by the main hypotheses? Similarly, could the authors comment on those task-based regressors predicting age/gender? Are they in line with the current literature?

2. Page 22. "95% % CI_lower:1.081, hypothesis 2C …" Do the authors mean hypothesis 2B here? As 2C refers to the hypothesis on RT.

3. Table 4. I suggest having a clear outcome report for each sub hypothesis. I.e., H1B/H2C is n/a nor not tested. H2A is not supported but H2B is confirmed (avoiding the ambiguous "partially confirmed").

4. Page 44 "Step 3: We applied the models from step to the confirmation sample" --> "from step 2" ?

[re PLOS' data availability policy]

The authors claimed that the data will be made available after acceptance. As a result, data availability/time stamps cannot be reviewed at this stage.

Reviewer #3:

[identifies himself as Christoph W. Korn]

In my view, Scholl and colleagues have provided a very comprehensive and detailed Stage 2 manuscript. Given this complexity, it took me longer than expected to carefully assess the criteria listed for Stage 2 manuscripts. I did not find any flaws. The few deviations from Stage 1 are explained (e.g., the RT analyses). In my view, only the last criterion is not yet fulfilled since the data is not yet freely accessible.

Below, I list a few minor suggestions for improving the clarity of the manuscript.

1. Figure 1F: Why does the model fit get worse with increasing numbers of previous searches? Is this simply because there are less data points available? Is there any particular reason for the "peak" at "6 previous searches" for "cost 15?"

2. In some places, the distinction between self-reported "pre-emptive avoidance" and the parameter 'AvgProspVChange' is a bit confusing.

a. At first, it sounds as if the authors just measured a self-report but it later becomes clear that they also have decision variable.

b. It was unclear to me why the authors did not look into 'AvgProspVChange' (see footnote e). This could be an interesting exploratory analysis.

c. In the abstract, the authors write: "In addition, this awareness allowed them to report pre-emptively avoiding situations related to the bias." The question that comes to mind immediately is: "Did participants actually pre-emptively avoid those situations?"

3. A few details could be mentioned briefly or explained more clearly in the results (and not "just" in the methods).

a. How high are the correlations between decision variables? (below .7).

b. The distinction in the analyses between the first and the later searches could be stressed a bit more in the results.

c. It could simply be stated that all results from the discovery sample are reported in the Supplemental Material and links to the corresponding figures/tables could be given.

4. Style:

a. Figure 3B (and others): The labels are difficult to read. Usually, the columns are ordered in the same way as the rows and therefore the main diagonal runs from upper left to lower right. Maybe, the easiest way is to just show the lower (or upper) triangular part of the matrix.

b. In the abstract (and also later), the term "perceived insensitivity" sounds ambiguous. "Self-reported" might be a better word.

c. In Table 4, "partially confirmed" could be replaced by a statement specifying the difference between the discovery and the confirmation samples.

d. Figure 1: The figure is quite complex and the caption is very long. Splitting this up could help.

e. There are some typos or omissions: e.g., "Batchelor," "large set regressor-to-factor relationships"

f. The title for the supplements is different.

Decision Letter 5

Roland G Roberts

3 Feb 2022

Dear Jacquie,

On behalf of my colleagues and the Academic Editor, Christopher Chambers, I'm pleased to say that we can in principle accept your Preregistered Research Article (our very first!) "The effect of apathy and compulsivity on planning and stopping in sequential decision making" for publication in PLOS Biology, provided you address any remaining formatting and reporting issues. These will be detailed in an email that will follow this letter and that you will usually receive within 2-3 business days, during which time no action is required from you. Please note that we will not be able to formally accept your manuscript and schedule it for publication until you have any requested changes.

Please take a minute to log into Editorial Manager at http://www.editorialmanager.com/pbiology/, click the "Update My Information" link at the top of the page, and update your user information to ensure an efficient production process.

PRESS: We frequently collaborate with press offices. If your institution or institutions have a press office, please notify them about your upcoming paper at this point, to enable them to help maximise its impact. If the press office is planning to promote your findings, we would be grateful if they could coordinate with biologypress@plos.org. If you have not yet opted out of the early version process, we ask that you notify us immediately of any press plans so that we may do so on your behalf.

We also ask that you take this opportunity to read our Embargo Policy regarding the discussion, promotion and media coverage of work that is yet to be published by PLOS. As your manuscript is not yet published, it is bound by the conditions of our Embargo Policy. Please be aware that this policy is in place both to ensure that any press coverage of your article is fully substantiated and to provide a direct link between such coverage and the published work. For full details of our Embargo Policy, please visit http://www.plos.org/about/media-inquiries/embargo-policy/.

Thank you again for choosing PLOS Biology for publication and supporting Open Access publishing. We look forward to publishing your study. 

Sincerely, 

Roli

Roland G Roberts, PhD 

Senior Editor 

PLOS Biology

rroberts@plos.org

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Text

    Contains Supporting information Methods, Results, and Supporting information Tables A–I. Table A: Parameter recovery (schedule used for confirmation sample). The table shows correlations (Pearson’s r) between simulated (“ground truth”) parameters (headers highlighted in grey) and the fitted (“recovered”) parameters (headers with white background) for the decision models of the initial (top) and later (bottom) searches. All parameters showed very good recovery (r > 0.88, diagonal, highlighted in blue). The off-diagonal values show that confusion between parameters was very low (all < 0.22). We used 500 simulated participants for these results. Data in files S13 and S14 Data. Table B: Parameter correlations for initial decisions (confirmation sample). We ran a Pearson correlation of the hierarchically fit parameter estimates from our cognitive model on the initial decisions to look at their relationships. Even though simulations (S1 Table) showed very low parameter confusions, parameters from real participants showed in parts substantial correlations (Pearson’s r), suggesting those correlations are a real relationship existing in participants, rather than an artifact of model fitting. invTemp (inverse temperature of the softmax equation linking value to choice probabilities), SearchBias (bias for or against searching). Prospective (prospective value of the model), Myopic (Myopic value of the model). AvgProspVChange (average change of prospective value per search in the future). Data in file S11 Data. Table C: Parameter correlations for later searches (confirmation sample). We ran a Pearson correlation of the hierarchically fit parameter estimates from our cognitive model on the later decisions to look at their relationships. Even though simulations (S1 Table) showed very low parameter confusions, parameters from real participants showed in parts substantial correlations (Pearson’s r), suggesting those correlations are a real relationship existing in participants, rather than an artifact of model fitting. Abbreviations same as S2 Table. ProspVS1 (prospective value at first search). ProspVS1mAdapted (change of prospective value since first search). PrevSearches (number of previous searches). Total Cost (sum of total cost since first search). Data in file S9 Data. Table D: Additional demographics (confirmation sample). Participants completed the following questionnaires. AMI [1], OCI-R [2], BDI [3], Spielberger State Anxiety [4], Liebowitz Social anxiety [5], Short Scales for Measuring Schizotypy [6] (“unusual experiences” and “introvertive anhedonia” subscale), Fatigue scale [7], and Toronto alexithymia scale [8]. Depression cutoffs were (less than 13 “Minimal,” less then 20 “Mild,” less than 29 “Moderate,” and higher or equal 29 “Severe”). Obsessive compulsion less than 21 “none,” otherwise “clinically significant”). For social anxiety, we had less than 30 as “none,” between 30 but less than 60 as “possible” and at and above 60 as “probable” social anxiety. For Alexithymia, we had less than 52 as none, between 52 but less than 61 as “possible” and at and above 62 as “present” Alexithymia. Fatigue was “none” below 22, “present” between 22 and 34, and “extreme” above 34. Data in file S5 Data. Table E: Results of regressions linking clinical dimensions to decision and self-report measures controlling for medication status (confirmation sample). As control regressors, we included age, gender and education (as in main manuscript). Additionally, here, psychoactive medication status is also included (coded as “Yes”/“No” as there were too few respondents of individual types of medications). Data in files S2, S6, S9, and S16 Data. Table F: Results of regressions linking clinical dimensions to later search decisions. Results of all parameters from model fit to the later search decisions and dimensions. We used 90% 2-sided CIs, to show the results of 1-sided (95% CI) tests for our preregistered hypotheses and to show what other links have any evidence. Data in files S9 and 16 Data. Table G: Results of regressions linking clinical dimensions to initial search decisions. Results of all parameters from model fit to the initial search decisions and dimensions. We used 90% 2-sided CIs, to show the results of 1-sided tests for our preregistered hypotheses and to show what other links have any evidence. Data in files S11 and S16 Data. Table H: Results of regressions linking clinical dimensions to self-report measures. Results of all self-report measures fit to the dimensions. We used 90% 2-sided CI, to show the results of 1-sided tests for our preregistered hypotheses and to show what other links have any evidence. Data in files S2 and S16 Data. Table I: Additional medication information (confirmation sample). All information for the types of medications excluded and included participants used (No and Yes). The most common medication were selective serotonin reuptake inhibitor. Data in file S6 Data. AMI, Apathy Motivation Index; BDI, Beck Depression Inventory; CI, confidence interval; OCI-R, Obsessive-Compulsive Inventory.

    (PDF)

    S1 Fig. Effects on decisions and deliberation time during initial decisions (confirmation sample).

    (A) Using a decision model, we tested which factors influenced the decision of whether to initiate a search. We found that participants were more likely to search when prospective and myopic values were high and that vice versa they were less likely to search when initial offer value or costs of searching were high. (B) We analysed the impact of different factors on RTs using a regression analysis. Of particular note, the larger the prospective value, the slower participants responded, potentially indicating more time taken to plan ahead. (C) When directly asked about their preemptive avoidance about half never reported using such strategies (self-report Q3). (E) Histogram of RTs across all participants, for initial searches on each trial (red) and later searches (blue). Note that for initial decisions there was a fixed delay of 3 to 6 seconds on each trial before participants could make responses, which is not included in the RT measured here. RTs for initial searches have an earlier mode than for later searches, but more slow RTs (“thicker tail” of the distribution). Error bars show Bayesian 95% credible intervals (2-tailed), significance is shown by credible intervals not including zero. Data of A in file 11, B in 12, C in 2, and D in 1. RT, reaction time.

    (TIFF)

    S2 Fig. Effects on decisions and deliberation time during later decision sequence (confirmation sample).

    (A) Once participants engaged with a sequence of searches, they had to decide how long to go on for. We found evidence for decision inertia (“# Prev. searches”) that was independent of effects of “sunk cost” fallacy (“Total cost”) or lack of updating of prospective value, i.e., their search strategy (“ProspVS1-Adapated”). Across all participants, we found sensitivity to costs (“Cost”). (B) RTs for later searches revealed that the more often participants had searched already within a given sequence, the faster they got (“#Prev. searches”). Moreover, as in the initial search (S1B Fig), the higher the cost, the slower they responded on these later searches in the sequence (“Cost”). (C) Model fits for the decision data, higher values indicate worse model fit (sum of log likelihoods across all participants and all trials, computed using cross-validation, relative to the best fitting model). The “full model”—the model used throughout the manuscript—provides the best fit to the data. The other models were derived from the “full model,” leaving out individual components of the model, one at a time. (D) Participants were also asked about their own task behaviour, such as their perceived cost insensitivity and most reported insensitivity to the costs to some extent (question only included in the confirmation sample). (E) Despite the strong choice effects of decision inertia, about half of the participants reported having no such bias (self-report Q 1). (F) Contrary to this, about 80% of people reported having been biased towards overchasing a rewarding option at least somewhat (self-report Q2). Error bars show Bayesian 95% credible intervals (2-tailed), significance is shown by credible intervals not including zero. Data of A in file 9, B in 10, C in 15, and D to F in 2.

    (TIFF)

    S3 Fig. Psychiatric symptom questionnaire results in discovery sample.

    (A) Histograms of the distribution of total questionnaire scores for the OCI-R (Ai), the AMI (Aii), the BDI (Aiii), and the Liebowitz social anxiety scale (Aiv). On each histogram, cutoffs based on previously published normative data are highlighted. (B) Correlation matrix (Pearson’s r) for all subscales included in the factor analysis. Squares highlight the factors subscales were assigned to in the factor analysis. (C) Loadings (i.e., the contribution of each subscale to each factor) of the factor analysis of the subscales. Highlighted in black are loadings above 0.4, for ease of visualisation. AMI, Apathy Motivation Index; BDI, Beck Depression Inventory; OCI-R, Obsessive-Compulsive Inventory.

    (TIFF)

    S4 Fig. Partial correlations between clinical factors and behavioural measures (confirmation sample), controlling for other clinical, demographic and general task performance measures (see Methods—“Relating questionnaires to task performance and self-reports”).

    Significance was assessed using nonparametric correlations (Kendall’s tau). Panel A corresponds to regression coefficients shown in Fig 4. Panels B–D correspond to regression coefficients shown in Fig 5 A–C. Note that for correlations of clinical factors with self-report questionnaires (C and D), responses needed to be treated as continuous, while in the regression analysis they were treated as ordered factors. Data of A to D in file 9, 2, and 16.

    (TIFF)

    S5 Fig. Parameter recovery (confirmation sample).

    We constructed 10 random schedules as detailed in the S1 Text section “Task schedule design and model validation.” Scatter plots show correlations between simulated and recovered parameters for the initial search (Ai) and later searches (Ai). Different colours show the different schedules (each with n = 500 simulated participants). Correlations (Pearson’s r) are shown across all 10 schedules. The average correlations (Pearson’s r) between regressors used in the RT regression and the decision-making models are at or below 0.7 for all variables for initial decisions (Bi) and later (Bii).” Data of A in file 13 and 14 and B in 1.

    (TIFF)

    S1 Data. Raw behavioural data (choices and RT) for each participant for each trial of the decision-making task.

    RT, reaction time.

    (CSV)

    S2 Data. Raw questionnaire answers (for each item of each questionnaire) and post-task self-reports for each participant.

    (CSV)

    S3 Data. Summary scores for questionnaire subscales (separately for each questionnaire) for each participant.

    See Fig 3B.

    (CSV)

    S4 Data. Demographics for each participant.

    See Table 1.

    (CSV)

    S5 Data. Clinical questionnaire summed scores for each participant.

    See Fig 3A and Table D in S1 Text.

    (CSV)

    S6 Data. Each participant’s medication status.

    See Table E and Table I in S1 Text.

    (CSV)

    S7 Data. Each participant’s probability of searching binned by relative value.

    See Fig 1E.

    (CSV)

    S8 Data. Each participant’s probability of searching binned by the number of previous searches and cost of searching.

    See Fig 1F.

    (CSV)

    S9 Data. Computational decision model parameters for each participant for data fitted on the decisions during the later searches on each trial.

    See Figs 2A, 4, and 5 as well as S2A and S4S6 Figs and Table 3.

    (CSV)

    S10 Data. Computational RT model parameters for each participant for data fitted on the decisions during the later searches on each trial.

    See S2B Fig and Table 3. RT, reaction time.

    (CSV)

    S11 Data. Computational decision model parameters for each participant for data fitted on the decisions during the initial search on each trial.

    See S1A Fig, Table 3, and Tables B and H in S1 Text.

    (CSV)

    S12 Data. Computational RT model parameters for each participant for data fitted on the decisions during the initial search on each trial.

    See S1B Fig and Table 3. RT, reaction time.

    (CSV)

    S13 Data. Decision model simulation: simulated (“true”) and recovered (“fitted”) parameters for later searches of each trial.

    See S5 Fig and Table A in S1 Text.

    (CSV)

    S14 Data. Decision model simulation: simulated (“true”) and recovered (“fitted”) parameters for the initial search of each trial.

    See S5 Fig and Table A in S1 Text.

    (CSV)

    S15 Data. Model fits (cross-validation) for each participant for the later searches of each trial.

    See S2C Fig.

    (CSV)

    S16 Data. Transdiagnostic dimensions (factor scores) for each participant.

    See Figs 3C, 4, and 5, S4 Fig, Tables 2 and 3 as well as Tables E–I in S1 Text.

    (CSV)

    Attachment

    Submitted filename: Reviewer_Letter_reducedFileSize.docx

    Attachment

    Submitted filename: Notes to the reviewers_250521.docx

    Attachment

    Submitted filename: Notes to the reviewers_250521.docx

    Attachment

    Submitted filename: Reply to the reviewers 170122.docx

    Data Availability Statement

    All anonymized data and analyses script are freely available on a repository (https://osf.io/dfg2u/). The data are also available in the supplementary data files (S1S16 Data).


    Articles from PLoS Biology are provided here courtesy of PLOS

    RESOURCES