Abstract
This paper responds to a recent critique by Bissett et al. of the fMRI Stop task used in the Adolescent Brain Cognitive Development℠ Study (ABCD Study®). The critique focuses primarily on a task design feature related to race model assumptions (i.e., that the Go and Stop processes are fully independent). In response, we note that the race model is quite robust against violations of its assumptions. Most importantly, while Bissett raises conceptual concerns with the task we focus here on analyzes of the task data and conclude that the concerns appear to have minimal impact on the neuroimaging data (the validity of which do not rely on race model assumptions) and have far less of an impact on the performance data than the critique suggests. We note that Bissett did not apply any performance-based exclusions to the data they analyzed, a number of the trial coding errors they flagged were already identified and corrected in ABCD annual data releases, a number of their secondary concerns reflect sensible design decisions and, indeed, their own computational modeling of the ABCD Stop task suggests the problems they identify have just a modest impact on the rank ordering of individual differences in subject performance.
Keywords: ABCD, STOP task, Race model, Adolescence, Neuroimaging
1. Introduction
Bissett et al. (2021) list a number of concerns regarding the specific version of the Stop Signal Task that is included as one of three fMRI tasks in the neuroimaging battery of the multi-site, longitudinal Adolescent Brain Cognitive Development (ABCD) study (www.ABCDstudy.org). The ABCD study is committed to full sharing of its tasks, data processing and analysis scripts, and dataset. We are delighted to see “open science” in action, and we value the scientific community flagging concerns and helping us to continually improve this landmark study.
The Bissett et al. critique focuses largely on stimulus design characteristics of the ABCD Stop Signal Task which may lead to violations of race model assumptions (Logan and Cowan, 1984). The concerns are well described in Bissett et al. In brief, the onset of the Stop signal, which is controlled by an adaptive, performance-related algorithm, is accompanied by the simultaneous offset of the Go choice stimulus. One consequence of this design feature is that the on-screen duration of the Go choice stimulus can be shorter on Stop trials. The concern raised by Bissett et al. is whether this might affect the context independence of the Go process, that is, that the Go process is unaffected by whether or not a Stop signal is presented on a trial. Context independence is assumed by race model theories of response inhibition and underpins the valid calculation of the Stop Signal Reaction Time (SSRT), an estimate of the duration of the stopping process. In addition to this primary issue, Bissett and co-authors raise a number of other concerns with the task design and conclude that these “significantly compromise” the value of the data. We disagree with this conclusion and provide evidence in support of the utility and validity of the Stop task data from the ABCD Study. The Bissett et al. critique raises important theoretical issues related to the assumptions of the race model underlying calculation of the SSRT, but the issues raised do not necessarily undermine the utility of the ABCD SSRT estimates as a measure of individual differences. Importantly, while Bissett et al. do raise conceptual concerns and do conduct a number of behavioral analyses (e.g., choice accuracy on STOP Fail trials) they do not examine empirically whether the issues they raised actually alter the SSRT estimates provided by the ABCD study, nor do they address whether they have any impact on the brain imaging data.
While acknowledging the concerns regarding race model violations, we focus here on empirically investigating the extent to which these violations meaningfully impact the quality of the ABCD data. We present a series of analyzes of both the SSRT estimates and the validity of the neuroimaging data. Ensuring independence of the Go and Stop processes is a perennial concern with the Stop task and not one peculiar to the ABCD task version as Bissett et al. have themselves argued (Bissett et al., 2021). Moreover, the Stop task is quite robust against violations of race model assumptions (Band et al., 2003). Thus, it is critical to determine whether, or to what degree, the concerns raised do indeed corrupt the SSRT estimates. We demonstrate here that they appear to have only modest effects on SSRT and, furthermore, do not substantively impact or invalidate the Stop task fMRI measures.
Some of the concerns raised by Bissett et al. were in fact known issues that were corrected prior to the annual data analyses and every curated data release of the ABCD Data Analysis, Informatics and Resources Center. Bissett et al. applied no performance flags (either their own or those recommended in all ABCD data releases) that serve to exclude participants who do not perform the task appropriately. Other recommended corrections, such as flipping left-right responses in those participants with reversed response paddles, were not applied. And, finally, we contend that many of the minor issues raised by Bissett et al. are design features and not design flaws. Following the format of the original paper, we address each issue in turn and offer our recommendations on if, and how, each might be addressed. The ABCD task fMRI working group in consultation with members of its External Scientific Board and outside experts have decided to make a number of small changes to the task. These changes are in consideration of the impact of the issues raised by Bissett et al. balanced against the implications of changing a task once a longitudinal study has commenced.
2. ABCD stop task
The code underlying the analyzes conducted in this paper are available (https://github.com/sahahn/SST_Response). We underscore that although the annual ABCD data releases provide updates and contain flags identifying problematic data (e.g., poor quality images, poor task performance), the data are fully available to the scientific community empowering researchers to apply whatever inclusion and exclusion criteria they deem appropriate for their specific research questions. Fig. 1 shows the performance criteria that have been used to calculate the Stop task performance flag that is part of the recommended inclusion criteria for these data provided with each ABCD data release. Fig. 1 shows two important properties of the ABCD Stop task data. First, performance statistics show that the tracking algorithm achieves a successful inhibition rate of approximately 50 % (which maximizes efficiency of SSRT estimation), few omission errors, and typically distributed response times. Second, it shows robust stopping-related activation in known response inhibition-related regions. These characteristics increase confidence that the ABCD Stop task data offer insightful measures of inhibitory control abilities and related brain function.
Fig. 1.
ABCD stop signal task. The performance criteria (recommendations for participant exclusion) and group level performance on the ABCD Stop Signal Task are shown in the tables. Histograms of the p(Successful Stopping) and SSRT and group activation maps (Cohen’s d threshold of .2 for Successful Stops vs Correct Go trials), for the baseline (age 9 and 10) data are shown.
2.1. Issue 1. Different go stimulus duration across trials
As noted above and described in Bissett et al., the offset of the Go choice stimulus coincident with the onset of the Stop signal may reduce the strength of the Go process on Stop trials, violating context independence. Poorer choice accuracy on Stop Fail trials compared to Go trials (Fig. 1), especially at shorter Stop Signal Delays, indicates that this is a valid concern; the poorer choice accuracy on the Stop Fail trials is consistent with the Go process being different on these trials in comparison to Go trials. In our analysis, with performance criteria applied and the exclusion of a small number of participants who experienced a task programming error (see Issue 3 below), Go accuracy = 91 % and Stop Fail accuracy = 81 %.2 In interpreting the poorer accuracy of Stop Fail trials, one consideration is that Stop Fail responses are faster than Go trial responses (by virtue of the full distribution of Stop trial responses being truncated by the Stop process winning the race) and faster responses are expected to be less accurate. For example, overall accuracy on Go trials is 91 % but Go accuracy on the fastest quintile of responses (< 378 ms) drops to 77 %. Choice accuracy for Stop Fail trials within this same range is 67 %. Combined, these results suggest that the faster responding of Stop Fail trials likely contributes to, but does not fully explain, their relatively lower accuracy. Beyond these speed-accuracy considerations, the critical issues are to what extent a violation of race model assumptions impacts reaction time and/or brain imaging data, and whether any violation impacts the utility of SSRT as a measure of individual differences in inhibitory control. Full context independence is difficult to attain. Bissett et al. and others demonstrate that this is the case across many Stop Signal Tasks, including those without the stimulus design feature of the ABCD task (Gulberti et al., 2014, Bissett et al., 2021 Mar 4). The presentation of a Stop signal has the potential to impact an ongoing Go response, even if the Go stimulus remains on screen. In addition, it is known that participants slow their Go responses in anticipation of a Stop trial (indeed, this “proactive” control can be modeled; Harlé et al., 2016; Logan et al., 2014). “Strategic” adjustments in the speed of Go trial responding can vary within an experiment and across individuals, and can affect SSRT estimates (Leotti and Wager, 2010).
One test for clear violations of context independence, which Bissett et al. presents, is whether Stop Fail response times (RTs) are slower than Go RTs. This pattern cannot be explained by the standard race model, which assumes Go processes are equivalent on Go and Stop signal trials and, further, that the observed distribution of RTs on Stop Fail trials are censored (by virtue of the Stop process completing before the relatively slower Go responses would be made). As Bissett et al. note, the ABCD data pass this test: Stop Fail trials are not slower than Go response times. Table 1 shows RT data for ABCD as calculated by Bissett et al., after we applied the standard performance flags, and excluded participants affected by the Issue 3 programming error (described below). Stop Fail RT is 83 ms faster than Go RT (t(1,7114) = 53, p < .0001), broadly consistent with context independence. For comparison, we include two datasets from Bissett et al., chosen because they do not share the ABCD design property of offset of the Go stimulus with the onset of the Stop stimulus and these datasets would be available to Bissett et al. for validation of our findings or for any further analyzes of their own. We refer to these datasets as the Ontology study (n = 522; Eisenberg et al., 2019) and the Phenome study (n = 130, healthy controls only; Poldrack et al., 2016). We use the low frequency condition (20 % Stop trials) in the Ontology study as it is closer to the ABCD proportion of 17 %. The ABCD value of 83 ms is within the range of these other studies: Stop Fail is 30 ms and 122 ms faster than Go trials in the Ontology and Phenome studies, respectively.
Table 1.
Performance statistics relevant to Issue 1 for ABCD and two comparison datasets (see text for details). We replicate the initial calculations by Bissett et al. of the mean response times for Go RT and Stop Fail (SF) RT trials, we apply the exclusion of participants based on poor performance as recommended with the ABCD annual data releases, and then we drop the RT data of an additional 1.24 % of participants whose task was compromised by a coding error (described under Issue 3 below).
ABCD |
Comparison |
||||
---|---|---|---|---|---|
Initial | Exclude poor performers | Issue 3 exclusion | Ontology | Phenome | |
Mean Go RT | 543 | 543 | 544 | 571 | 478 |
Mean SF RT | 459 | 458 | 460 | 541 | 356 |
Difference | 84 | 85 | 83 | 30 | 122 |
Next, Bissett et al. reports that 6.2 % of the ABCD participants do not show the expected RT pattern and label them as violators (i.e., participants for whom Stop Fail RT > Go RT). While one might refine this estimate (it drops to 5 % once performance flagged participants and participants on whom there was a task programming error described below are excluded) it is nonetheless superior or comparable to the estimates for the Ontology and Phenome studies (17 % and 4.4 %, respectively) which we again note did not share the stimulus design features of the ABCD Stop task. These percentages may not indicate statistically reliable effects (i.e., the differences in RTs may not differ reliably from 0), as confidence intervals and estimates of their sampling distribution are not provided. Although analyzes of individual participants are likely underpowered, with too few trials for robust behavioral analyses, one can estimate the numbers of participants who might be deemed true violators (i.e., with a significant one-tailed t-test per participant showing Stop Fail trial RTs to be longer than Go trial RTs). The percentage is low: 1.6 % for ABCD, again falling between the Ontology (2.7 %) and Phenome (0 %) studies that were designed without coincident Go stimulus offset and Stop stimulus onset.
We conclude that some number of participants evidencing context independence violations are to be expected in many Stop task designs. While researchers can make their own decisions on participant inclusion and exclusion criteria given ABCD’s data sharing procedures, we note that a recent Stop task “best practices” paper recommends that SSRT should not be estimated for those participants who violate the Stop Fail RT < Go RT criterion (Verbruggen et al., 2019). Consequently, a new flag identifying the 5 % of ABCD participant “violators” is included in the annual ABCD data releases (see Implications and Recommendations below). Note that this flag is applied to any participant whose Stop Fall RT > Go RT by any amount (i.e., 1 ms or greater).
2.1.1. Individual differences
Examining individual differences is a central goal of the ABCD study. Crucially, for the utility of the ABCD SSRT estimate to be degraded as a measure of individual differences, violations of context independence must result in more than a shift in mean SSRT. Rather, the rank ordering of participants’ SSRT values must be substantively altered. As all ABCD participants performed the same task, we expect individual differences to be largely unaffected by the abbreviation of the Go process as a function of task structure, but this expectation awaits further experimental or computational modeling studies. Computational modeling approaches have potential to characterize the specific design features of the ABCD Stop task and the additional processes that can occur on all Stop tasks (e.g., “trigger errors”, or trials on which the STOP process is never initiated; Weigard et al., 2019). Bissett et al. describe some preliminary drift diffusion models to capture the particulars of the ABCD task design. While we leave it to others to judge the value of these models, we note that they return “adjusted” SSRT estimates that don’t, in fact, appear to be that different to the estimates derived from the standard analyses that assumes context independence. To elaborate, an essential element of the Bissett et al. model is the estimate of inter-individual variation in SSRT. To assign an SSRT value to their simulated subjects, they “sampled randomly from an SSRT distribution with a mean that equaled the observed ABCD grand mean but assumed four different amounts of between-subject variability (ranging from SD = 0–85 ms).” Their own analyzes of “20 simple stopping conditions from a recent large-scale stopping study” estimated the mean between-subject SD of SSRT to be 43 ms with a range of 28–85 ms. At the higher estimate (85 ms) they estimate the mean rank correlation between SSRT as calculated by the Independent Race Model (i.e., assuming context independence) and their three alternative models to be.93. Unfortunately, they did not report the correlation for 43 ms although this is the empirical mean that they estimated from their own analyzes of their 20 simple stopping conditions. The minimal estimate from their analyzes is 28 ms but the correlation for this value is also not reported. Instead, the nearest estimate to the bottom of their observed range (25 ms) yields a mean correlation of .78. The other estimates that they report in their paper, and to which they pay specific attention, are 5 ms and 0 ms. However, these are far beyond the range of their empirically observed estimates and are not at all credible estimates of the true inter-subject variability in SSRT. Using their code, we have repeated their analyzes using their mean estimate of SSRT SD (43 ms). We observe a mean correlation between their computational models and the ideal independence race model to be .85. The true estimate of inter-subject variability in SSRT in the ABCD data is, of course, uncertain if one holds that these SSRT estimates are invalid. Nonetheless, if we exclude violators, remove participants who experienced a task programming error (see Issue 3 below), exclude participants flagged for poor performance, and calculate SSRT with the 0 ms SSD trials excluded (see Issue 2 below), we calculate the SSRT SD to be 73 ms. This estimate, in turn, yields a mean correlation between their computational models and the ideal independence race model to be .91. Thus, from the simulations and computational modeling conducted by Bissett et al., we conclude that modeling the context violation present in the ABCD study is unlikely to distort the rank ordering of participants in a meaningful way.
Future experimental or computational modeling studies could help provide greater clarity about these matters. Until that time, researchers are encouraged to carefully consider the assumptions of any measurement model they apply to these data, including the race model-based SSRT estimate, and to consider the possible limitations of parameter estimates and measures derived from any model that assumes context independence.
2.1.2. Stop task brain activation
Turning to the brain activation data, it is important to note that the measurement assumptions that the activation contrasts reflect valid measures of response inhibition are much simpler than those required for the SSRT estimation and do not rely on race model assumptions, including context independence. Brain activity can be associated with response inhibition processes if one compares trials requiring inhibition of prepotent responses against trials that do not. A standard contrast to achieve this end for ABCD would be to compare Successful Stop trials against Go trials. The shorter duration of Go choice stimuli on Stop trials compared to Go trials does introduce differences between the two conditions. However, there are typically other more substantial differences present when isolating inhibition-related activation: There is a motor response on Go trials and not on Stop success trials, only the latter contains a Stop signal, and so on. The contrast of Successful Stop trials against the implicit baseline and the contrast of Successful Stop trials against Failed Stop trials are also available to researchers. As is always the case, researchers using these data should be aware of design specifics and determine if they impact on the researcher’s specific question. The ABCD Stop task has already been shown to produce robust activation in the response inhibition network and activation levels show the anticipated correlations with individual differences in SSRT (Casey et al., 2018, Chaarani et al., 2021). A very similar task, with the same Go stimulus design features, has been employed in the IMAGEN study of adolescent development (Schumann et al., 2010) and has, for example, identified functional differences between adolescents with substance use, adolescents with ADHD, adolescents with psychotic symptoms, dysregulated youth and controls (Bourque et al., 2017; Spechler et al., 2019a; Whelan et al., 2012), and has predicted future drug use (Spechler et al., 2019b, Whelan et al., 2014).
We demonstrate the validity of the brain activation measures in ABCD with two analyzes. The first examined brain activation in the violators identified above (a “worst case” scenario in which we might expect atypical activation patterns) and the second examined the impact of removing trials with short Stop Signal Delays (SSD) from all participants. Short SSD trials (SSD < 150 ms), in which the Go stimuli were presented for the shortest durations, are likely to be the trials driving any context independence violations. The first analysis compared Stop activation (the contrasts Successful Stops and Successful Stops vs Correct Go trials) between the group of violators (n = 257; baseline [ages 9 and 10] data only) and non-violators (n = 5001). Covariates included sex, age (in months), highest parental education, race/ethnicity, puberty level, and scanner. No group differences were observed with a vertex-wise threshold of p < .05 and application of either a family-wise error correction based on random field theory or a less conservative false discovery rate. At a nominal p < .05, uncorrected threshold, group differences were observed in visual cortex only. Further, in a region-of-interest analysis of the right IFG, a critical node of the response inhibition network, violators and non-violators did not differ in activation (p = .83).
To quantify similarity between groups in the spatial patterns of activation, we calculated the vertex-wise correlation between group activation of the violators and a “gold standard” of activation based on the remaining non-violator participants (the whole sample was first residualized for the covariates listed above). To facilitate interpretation, we quantified the similarities that would be expected with samples of size 257 by comparing randomly selected subsamples of non-violators (n = 257, 10,000 samples) against the remainder (n = 4487; Fig. 2A). Fig. 2b shows the distribution of vertex-wise correlations for both the Successful Stops contrast (blue) and the Successful Stops vs Correct Go contrast (orange). The correlations for violators are indicated by the red line. Although lower than the mean of the subsampling distribution, shown by the dotted blue line, we note that the vertex-wise correlation is very high, even for violators (r = 0.94 and.94 for violators, compared to the full group of non-violators with r = 0.95 and.96, for Successful Stops and Successful Stops vs Correct Go, respectively). Moreover, violators, unsurprisingly, show relatively poor performance on the task (Fig. 2c). An equal sized group of non-violators, matched to the violators on Stop success rate, Stop Fail accuracy, Go RT, Go accuracy, and Go omission rate are indicated by the green lines (Matched Group). We conclude that even those participants identified as violating race model assumptions show activation patterns that are very similar to those observed for non-violators.
Fig. 2.
Correlation of violators and performance-matched non-violators with gold-standard brain activation. (a) A subsampling procedure determined the similarity between vertex-wise activation levels in samples of n = 257. (b) Correlations with the “gold standard” activation map for subsamples of non-violators (blue and orange distributions), violators (red line) and performance-matched non-violators (green line). (c) Performance of violators, all non-violators, and performance-matched non-violators.
The second analysis compared group Stop activation maps (Successful Stops vs Correct Go) with all trials vs with the shorter SSD Stop trials (0 ms, 50 ms, 100 ms) excluded. The same covariates as described above were included (n = 5058). Although the amplitude of activation was larger in the former (due, presumably, to the inclusion of more trials), critically, the patterns of activation were almost identical (Fig. 3). The vertex-wise correlation between the group activation map that included all trials with the group activation map that excluded the 0 ms, 50 ms, and 100 ms SSD trials was r = 0.99.
Fig. 3.
Brain activation (Cohen’s d threshold of.2 for Successful Stops vs Correct Go trials) for all Stop trials, Stop trials with SSD = 0/50/100 ms excluded, Stop trials with SSD = 0 ms excluded, and first ten Stop trials excluded.
2.1.3. Issue 1 implications and recommendations
Race model violators (Stop Fail RT > Go RT) are now identified in annual data releases and we suggest that researchers not include them when estimating SSRT (Verbruggen et al., 2019). We do not believe that there is sufficient evidence at this stage to warrant distrust of the remaining performance and neuroimaging data but encourage investigators to consider the impact of the issues that have been raised for their research question and their application of SSRT measurement models, and to apply what they deem to be appropriate inclusion and exclusion criteria. The ABCD task fMRI working group, in consultation with its External Scientific Board and additional external experts, have decided that changes to the fundamental task design are not warranted at this stage. The group notes that the poorer choice accuracy on Stop Fail trials does indicate a degree of independence violation that likely would be reduced if the Go stimuli remain on screen (for the 1 s duration of the trial or until a response is made). However, noting that context independence violations are common across numerous task designs (Bissett et al., 2021), the analyzes reported above suggest that the current task is yielding valuable, useful data that would appear to be largely unaffected by the concerns raised by Bissett et al. Clearly, an ongoing longitudinal study will prioritize not changing a task without compelling, and ideally empirical, reasons to do so.
2.2. Issue 2. Go stimulus sometimes not presented
Arising from the stimulus design feature underlying Issue 1, Go choice stimuli are not presented on trials in which the SSD drops to 0 ms. Bissett et al. suggest that this may confuse participants or “may make successfully stopping trivial (as the go process never started).” Bissett et al. report that 9.1 % of Stop trials are 0 ms SSD trials but once the performance flagged participants and the Issue 3 programming error participants are excluded this drops to 6.9 %. Curiously, performance on these trials is not trivial (average successful inhibition rate across participants on these trials is 64.8 %) reflecting, presumably, the response prepotency induced by the high proportion of Go trials (see Table 2) which suggests that these trials also engaged inhibitory control processes.
Table 2.
The number of Stop trials in which the SSD = 0 ms and the probability of successfully inhibiting a response.
All Ss (%) | Exclude poor performers (%) | Issue 3 exclusion (%) |
|
---|---|---|---|
% of trials | 9.1 | 7.3 | 6.9 |
Stop Success Rate | 60.3 | 63.7 | 64.8 |
The 0 ms SSD trials are broadly distributed across participants, with 49.9 % of participants having at least one 0 ms SSD trials. We calculated SSRT with these 0 ms SSD trials included (283 ± 78) and excluded (276 ± 73). (The average of a participant’s SSDs is included in the calculation of their SSRT, so the exclusion of all 0 ms SSD trials would be expected to produce a shorter SSRT estimate.) Notably, the correlation between the two estimates is very high (r = 0.97). The broad distribution of these trials across participants appears to reduce their impact on subsequent analyses. Turning to the neuroimaging data, the vertex-wise correlation between brain activation when these trials are included vs excluded is very high (r = 0.99; see Fig. 3). This correlation holds when all participants are included (n = 5064) and when the analyzes are restricted to those participants with one or more 0 ms SSD Stop trials (n = 2416).
2.2.1. Issue 2 implications and recommendations
To avoid potential confusion, the task has been modified to ensure that the SSD does not drop below 50 ms thereby ensuring presentation of the Go stimulus on all trials. The impact of the 0 SSD trials on the existing data appears to be very small. For researchers who may wish to exclude participants with a high number of these trials, we now include the number of 0 ms SSD trials per participant in annual data releases.
2.3. Issue 3. Degenerate stop-signal delays
Bissett et al. identified a programming error and we thank them for bringing this to our attention. When the SSD is 50 ms, a response that is faster than 50 ms is erroneously recorded as the response for all subsequent Stop trials. Bissett et al. report that this programming error affects 2.67 % of participants. However, if the performance flags are applied this reduces to 1.24 % of participants. The data of many of these participants are likely retrievable if one restricts analyzes to the data obtained prior to the onset of the error. For example, just 0.8 % of participants have this problem occur prior to their 50th Stop trial and the “best practices” paper by Verbruggen et al. concludes from a series of simulations that “reliable and unbiased SSRT group-level estimates can be obtained with 50 stop trials.” For participants with complete data and no glitch, we calculated SSRT based on all trials and again based on trials up to their 50th Stop trial and found highly consistent estimates (r = 0.97). While some caution might be warranted in assuming that this holds for participants who experienced the glitch (these participants have relatively long SSRTs) these results suggest that valid data may be obtained from the first 50 Stop trials of participants.
2.3.1. Issue 3 implications and recommendations
The programming error in the task has been corrected and the corrected task is available on the ABCD study website (https://abcdstudy.org/families/abcd-fmri-tasks-and-tools/). For the existing data, although we anticipate that valuable performance and brain imaging data of many participants affected by this programming error are retrievable, we include a variable identifying these participants in data releases, enabling researchers to exclude them from analyzes.
2.4. Issue 4. Different Stop Signal duration for different SSDs
This issue arose because all trial events were constrained to happen within the 1 s trial period and not to carry into the inter-trial interval. As a consequence, the duration of the Stop signal (typically 300 ms) was shortened if the SSD was greater than 700 ms. One might expect these events to be quite uncommon and also to be indicative of other performance-related problems (i.e., a 700 ms SSD is atypically long). We calculate the frequency of these shorter Stop signal durations to be 1.15 % of all Stop trials. However, this reduces to 0.12 % of trials once performance criteria are applied (as mentioned, the presence of these very long SSD trials indicates other performance problems). Moreover, as participants very often respond during the relatively long SSD, the proportion of trials in which the shorter SSDs are presented and on which participants have not already responded reduces to 0.07 % of trials. We removed these short Stop signal duration trials and assessed the impact on SSRT. Identical results were obtained: With all trials included and with these short Stop signal duration trials included, SSRT = 283 ± 78.
2.4.1. Issue 4 implications and recommendations
Given the rarity of these trials, we conclude that they have a negligible impact on the SSRT estimates. Nonetheless, we have changed the task to ensure that the Stop signal duration is always 300 ms in duration, a change which we believe will have a negligible impact on any longitudinal comparisons.
2.5. Issue 5. Non-uniform conditional trial probabilities
Bissett et al. note correctly that the trial orders (Stop vs Go) were not fully randomized (in accordance with the ratio of Stop to Go trials). Instead, the conditional trial probabilities and inter-stimulus intervals (ISIs) for ABCD were selected to optimize the joint estimation efficiency for fMRI responses to Go and Stop events, assuming a canonical hemodynamic response function. For each of 40 runs, sequences of Stop and Go stimuli, and randomized jittered intervals between them, were selected by generating 100,000 random trial sequences and ISIs (constrained to 0.7–2 s, starting mean 0.9 s), and choosing the design with the minimum mean Variance Inflation Factor. The design was constrained to avoid repeating successive Stop trials. The rationale for this is based on evidence that repeated Stops are easier, thereby reducing power and homogeneity in the demand on inhibition across trials (Bissett and Logan, 2012).
Bissett et al. raise concerns that the non-uniform trial probabilities might be learned by participants. Their analysis of post-Stop RTs indicates that this is not the case (i.e., initial post-Stop slowing does not transition, as the task progresses, to post-Stop speeding up) but they caution that it may be learned as the participants age. A very similar Stop task (i.e., no repeated Stop trials) has been used in the IMAGEN longitudinal study of ~ 2000 adolescents. We assessed post-Stop behavior similar to Bissett et al. (post-Stop RT minus pre-Stop RT, separating all the Stop trials into quartiles wherein the first quartile refers to the first quarter of trials across the task’s two runs, the second quartile refers to the second quarter of trials, and so on) when these participants performed the task at age 19, which was the second administration of the task to them. Fig. 4 shows no evidence that participants at age 19 learn to speed up on post-Stop trials as the task proceeds.
Fig. 4.
Boxplots showing response times on Go trials that immediately precede STOP trials and changes in response times for trials immediately following Stop trials (i.e., post-Stop RT minus pre-Stop RT). RTs are shown across trial quartiles for the IMAGEN participants at age 19. This task also excludes repeated Stop trials and, similar to the younger ABCD participants, shows no evidence of post-Stop speeding between the start and end of the task.
2.5.1. Issue 5 implications and recommendations
We recommend no changes to the task design. Researchers concerned about trial conditional probabilities being learned as the ABCD participants age can assess these in a manner as suggested by Bissett et al. Indeed, the extent to which participants develop subjective expectations and prepare for Stop trials is an interesting avenue of research for investigators interested in proactive control.
2.6. Issue 6. Trial accuracy incorrectly coded
Bissett et al. describe a number of trial outcome labeling errors. Unfortunately, Bissett et al. did not explain that all of these errors had already been identified by the ABCD Data Analysis and Informatics Resource Center, that they were corrected and documented prior to data analysis and data release, and that the code to make these corrections was shared publicly. These errors derived from how E-Prime pre-release settings were used: Responses occurring during the pre-release period of the post-trial jittered intervals were not recorded correctly, resulting in Stop Fail trials being logged as Stop Successes and Correct Go trials being logged as Go Omissions. Details on the trial labeling errors and their correction are included in the data processing scripts available from the ABCD github site (https://github.com/ABCD-STUDY/). The misclassification of errors on the Stop trials (classifying what should be Stop Fails as Stop successes) did impact the task (i.e., SSD increased when it should have decreased), but this occurred on only a small fraction of trials and, as Bissett et al. report, would have had a very small effect on the SSD tracking algorithm.
2.6.1. Issue 6 implications and recommendations
The cause of the errors has been corrected in the task code. Data available through the annual NDA releases have already corrected for these labeling errors but researchers wishing to work with the raw E-Prime output are encouraged to employ the corrections specified in the “abcd_extract_eprime_sst” script available on the ABCD github site.
2.7. Issue 7. SSD values start too short
Bissett et al. query the reasoning behind starting the task with the SSD set to 50 ms rather than a value that is closer to what the final SSD would be (e.g., 250 ms). The shorter SSD starts the task at a relatively easy level thereby providing a “warm-up” of sorts, easing participants, aged just 9 and 10 at baseline, into what is a cognitively challenging task. As shown in Figure 9 of Bissett et al., these first few trials are, in fact, not trivial for participants insofar as performance is better than average but far from ceiling (starting at 75 % and dropping to 60 % by the fifth Stop trial). Consequently, these first few trials contribute usefully to the behavioral and activation measures. As shown in the group performance table in Fig. 1, the 60 Stop trials of the ABCD task are sufficient for the adaptive algorithm to converge on ~ 50 % Stop success rate. To assess the impact of these starting trials, we removed the first ten Stop trials and all Go trials up to the tenth Stop trial and assessed the impact on SSRT. With all trials included SSRT = 283 ± 78 and with the first ten Stop trials excluded SSRT = 284 ± 85 with a high correlation between the two (r = 0.98). Similarly, group brain activation with and without these first ten Stop trials was calculated (n = 5067) with the vertex-wise correlation between the two being very high, r = 0.99 (see Fig. 3).
2.7.1. Issue 7 implications and recommendations
We believe there are no implications for the task or the data arising from this design feature.
2.8. Issue 8. Low stop trial probability
Bissett et al. raise concerns with the low stop trial probability in ABCD (17 %; the ratio of Go to Stop trials is 300:60). We note that the task contains 60 Stop trials (as mentioned above, 50 are deemed sufficient for estimating SSRT) and successfully converges on ~ 50 % Stop success rate (Fig. 1). As noted above, a very similar task, with the same Stop trial probability has been used in one of the largest adolescent neuroimaging longitudinal studies (IMAGEN; n = 2223, with assessments at ages 14, 19, and 23; Schumann et al., 2010) and it has demonstrated robust Stop success and Stop fail activations at all three ages and has demonstrated its ability to discriminate among adolescent phenotypes and predict future adolescent behavior (Whelan et al., 2012, Whelan et al., 2014). Importantly, a body of literature shows that estimates of SSRT are not affected by the proportion of Stop trials (Bissett and Logan, 2011, Logan and Burkell, 1986, Logan, 1981).
2.8.1. Issue 8 implications and recommendations
We believe there are no implications for the task or the data arising from this design feature. If researchers wish to compare the task’s activation levels or SSRT estimates against other datasets then, as is always the case in comparing separate studies, they should be aware of the particular design features of the ABCD study.
Conclusion We encourage researchers to continue analyzing the ABCD data and to help us improve the study by identifying potential problems like Bissett et al. have done. In light of the concerns raised by Bissett et al., some changes to the ABCD Stop task have been made (Table 3). Quantifying the impact of the specific concerns on the validity of the data is, of course, challenging: Although a specific cause for context violations can be identified in the ABCD Stop task, context violations as Bissett et al. have noted elsewhere (Bissett et al., 2021) are, in fact, quite widespread across multiple task designs and, consequently, researchers must always attend to this and other assumptions of their measurement models and analyzes. The context independence violation concerns are most pronounced on trials with shorter SSDs so likely have small effects on the majority of trials for the majority of participants and can be mitigated further by some of our recommendations such as deleting the 0 SSD trials. In total, the analyzes presented here lead us to conclude that the specific design feature of the ABCD Stop task appears, thus far, to have a minimal impact on the neuroimaging data. The impact on the SSRT data, including on the rank ordering of participants, appears to be modest, especially if the recommendations provided here (which does include some new participant exclusions) are followed. That said, we await more empirical and computational analyzes on these matters and encourage researchers to consider the implications of the task design for the analyzes they conduct and any measurement model they apply to these data. More generally, we encourage researchers to contact the ABCD team promptly should their analyzes raise concerns with any element of the assessment battery. Doing so ensures that misunderstandings can be avoided and any errors speedily corrected.
Table 3.
Recommended task and data sharing changes.
|
Data Statement
ABCD data are available through the NIMH Data Archive and the code underlying the analyzes conducted in this paper are available at https://github.com/sahahn/SST_Response.
Declaration of Competing Interest
The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
Acknowledgements
We thank Harriet de Wit, Ted Satterthwaite, Alex Weigard, members of the ABCD External Scientific Board, and members of the ABCD task fMRI working group for their helpful discussions of these topics. We thank Gordon Logan and an anonymous reviewer for their thoughtful and constructive review of this manuscript. Funding support: NIDA U01DA051039.
Footnotes
Excluding participants who fail to adequately comply with task instructions or who show aberrant performance is a standard practice and especially important given the young age of ABCD participants. Recommended performance criteria that accompany ABCD data releases are shown in Fig. 1 but these were not applied by Bissett et al. In addition, a programming error detailed under Issue 3 led to a very different task experience for certain participants and we are recommending that these participants (1.24 % of the sample) be excluded from analyzes.
Data Availability
ABCD data are available through the NIMH Data Archive and the code underlying the analyzes conducted in this paper are available at https://github.com/sahahn/SST_Response.
References
- Band G.P.H., van der Molen M.W., Logan G.D. Horse-race model simulations of the stop-signal procedure. Acta Psychol. 2003;112:105–142. doi: 10.1016/s0001-6918(02)00079-3. [DOI] [PubMed] [Google Scholar]
- Bissett P.G., Logan G.D. Balancing cognitive demands: control adjustments in the stop-signal paradigm. J. Exp. Psychol. Learn. Mem. Cogn. 2011;37(2):392–404. doi: 10.1037/a0021800. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bissett P.G., Logan G.D. Post-stop-signal adjustments: inhibition improves subsequent inhibition. J. Exp. Psychol. Learn. Mem. Cogn. 2012;38(4):955–966. doi: 10.1037/a0026778. [DOI] [PubMed] [Google Scholar]
- Bissett P.G., Jones H.M., Poldrack R.A., Logan G.D. Severe violations of independence in response inhibition tasks. Sci. Adv. 2021;7(12):eabf4355. doi: 10.1126/sciadv.abf4355. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bissett P.G., Hagen M.P., Jones H.M., Poldrack R.A. Design issues and solutions for stop-signal data from the Adolescent Brain Cognitive Development (ABCD) study. Elife. 2021;10 doi: 10.7554/eLife.60185. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bourque J., Spechler P.A., Potvin S., et al. Functional neuroimaging predictors of self-reported psychotic symptoms in adolescents. Am. J. Psychiatry. 2017;174(6):566–575. doi: 10.1176/appi.ajp.2017.16080897. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Casey B.J., Cannonier T., Conley M.I., Cohen A.O., Barch D.M., Heitzeg M.M., Soules M.E., Teslovich T., Dellarco D.V., Garavan H., Orr C.A., Wager T.D., Banich M.T., Speer N.K., Sutherland M.T., Riedel M.C., Dick A.S., Bjork J.M., Thomas K.M., Chaarani B., Mejia M.H., Hagler D.J., Jr., Daniela Cornejo M., Sicat C.S., Harms M.P., Dosenbach N.U.F., Rosenberg M., Earl E., Bartsch H., Watts R., Polimeni J.R., Kuperman J.M., Fair D.A., Dale A.M., Imaging Acquisition Workgroup A.B.C.D. The ABCD study: functional imaging acquisition across 21 sites. Dev. Cogn. Neurosci. 2018;32:43–54. doi: 10.1016/j.dcn.2018.03.001. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Chaarani B., Hahn S., Allgaier N., Adise S., Owens M.M., Juliano A.C., Yuan D.K., Loso H., Ivanciu A., Albaugh M.D., Dumas J., Mackey S., Laurent J., Ivanova M., Hagler D.J., Cornejo M.D., Hatton S., Agrawal A., Aguinaldo L., Ahonen L., Aklin W., Anokhin A.P., Arroyo J., Avenevoli S., Babcock D., Bagot K., Baker F.C., Banich M.T., Barch D.M., Bartsch H., Baskin-Sommers A., Bjork J.M., Blachman-Demner D., Bloch M., Bogdan R., Bookheimer S.Y., Breslin F., Brown S., Calabro F.J., Calhoun V., Casey B.J., Chang L., Clark D.B., Cloak C., Constable R.T., Constable K., Corley R., Cottler L.B., Coxe S., Dagher R.K., Dale A.M., Dapretto M., Delcarmen-Wiggins R., Dick A.S., Do E.K., Dosenbach N.U.F., Dowling G.J., Edwards S., Ernst T.M., Fair D.A., Fan C.C., Feczko E., Feldstein-Ewing S.W., Florsheim P., Foxe J.J., Freedman E.G., Friedman N.P., Friedman-Hill S., Fuemmeler B.F., Galvan A., Gee D.G., Giedd J., Glantz M., Glaser P., Godino J., Gonzalez M., Gonzalez R., Grant S., Gray K.M., Haist F., Harms M.P., Hawes S., Heath A.C., Heeringa S., Heitzeg M.M., Hermosillo R., Herting M.M., Hettema J.M., Hewitt J.K., Heyser C., Hoffman E., Howlett K., Huber R.S., Huestis M.A., Hyde L.W., Iacono W.G., Infante M.A., Irfanoglu O., Isaiah A., Iyengar S., Jacobus J., James R., Jean-Francois B., Jernigan T., Karcher N.R., Kaufman A., Kelley B., Kit B., Ksinan A., Kuperman J., Laird A.R., Larson C., LeBlanc K., Lessov-Schlagger C., Lever N., Lewis D.A., Lisdahl K., Little A.R., Lopez M., Luciana M., Luna B., Madden P.A., Maes H.H., Makowski C., Marshall A.T., Mason M.J., Matochik J., McCandliss B.D., McGlade E., Montoya I., Morgan G., Morris A., Mulford C., Murray P., Nagel B.J., Neale M.C., Neigh G., Nencka A., Noronha A., Nixon S.J., Palmer C.E., Pariyadath V., Paulus M.P., Pelham W.E., Pfefferbaum D., Pierpaoli C., Prescot A., Prouty D., Puttler L.I., Rajapaske N., Rapuano K.M., Reeves G., Renshaw P.F., Riedel M.C., Heath P., de la Rosa M., Rosenberg M.D., Ross M.J., Sanchez M., Schirda C., Schloesser D., Schulenberg J., Sher K.J., Sheth C., Shilling P.D., Simmons W.K., Sowell E.R., Speer N., Spittel M., Squeglia L.M., Sripada C., Steinberg J., Striley C., Sutherland M.T., Tanabe J., Tapert S.F., Thompson W., Tomko R.L., Uban K.A., Vrieze S., Wade N.E., Watts R., Weiss S., Wiens B.A., Williams O.D., Wilbur A., Wing D., Wolff-Hughes D., Yang R., Yurgelun-Todd D.A., Zucker R.A., Potter A., Garavan H.P., ABCD Consortium Baseline brain function in the preadolescents of the ABCD Study. Nat. Neurosci. 2021;24(8):1176–1186. doi: 10.1038/s41593-021-00867-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Eisenberg I.W., Bissett P.G., Zeynep Enkavi A., et al. Uncovering the structure of self-regulation through data-driven ontology discovery. Nat. Commun. 2019;10(1):2319. doi: 10.1038/s41467-019-10301-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gulberti A., Arndt P.A., Colonius H. Stopping eyes and hands: evidence for non-independence of stop and go processes and for a separation of central and peripheral inhibition. Front. Hum. Neurosci. 2014;8:61. doi: 10.3389/fnhum.2014.00061. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Harlé K.M., Zhang S., Ma N., Yu A.J., Paulus M.P. Reduced neural recruitment for bayesian adjustment of inhibitory control in methamphetamine dependence. Biol. Psychiatry Cogn. Neurosci. Neuroimaging. 2016;1(5):448–459. doi: 10.1016/j.bpsc.2016.06.008. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Leotti L.A., Wager T.D. Motivational influences on response inhibition measures. J. Exp. Psychol.: Hum. Percept. Perform. 2010;36(2):430–447. doi: 10.1037/a0016802. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Logan G.D. In: Attention and Performance IX. Long J., Baddeley A.D., editors. Erlbaum; Hillsdale: 1981. Attention, automaticity, and the ability to stop a speeded choice response; pp. 205–222. [Google Scholar]
- Logan G.D., Burkell J. Dependence and independence in responding to double stimulation: a comparison of stop, change, and dual-task paradigms. J. Exp. Psychol.: Hum. Percept. Perform. 1986;12(4):549–563. [Google Scholar]
- Logan G.D., Cowan W.B. On the ability to inhibit thought and action: a theory of an act of control. Psychol. Rev. 1984;91:295–327. doi: 10.1037/a0035230. [DOI] [PubMed] [Google Scholar]
- Logan G.D., Van Zandt T., Verbruggen F., Wagenmakers E.J. On the ability to inhibit thought and action: general and special theories of an act of control. Psychol. Rev. 2014;121(1):66–95. doi: 10.1037/a0035230. [DOI] [PubMed] [Google Scholar]
- Poldrack R.A., Congdon E., Triplett W., et al. A phenome-wide examination of neural and cognitive function. Sci. Data. 2016;3 doi: 10.1038/sdata.2016.110. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Schumann G., Loth E., Banaschewski T., et al. The IMAGEN study: reinforcement-related behaviour in normal brain function and psychopathology. Mol. Psychiatry. 2010;15(12):1128–1139. doi: 10.1038/mp.2010.4. [DOI] [PubMed] [Google Scholar]
- Spechler P.A., Chaarani B., Orr C., et al. Neuroimaging evidence for right orbitofrontal cortex differences in adolescents with emotional and behavioral dysregulation. J. Am. Acad. Child Adolesc. Psychiatry. 2019;58(11):1092–1103. doi: 10.1016/j.jaac.2019.01.021. [DOI] [PubMed] [Google Scholar]
- Spechler P.A., Allgaier N., Chaarani B., et al. The initiation of cannabis use in adolescence is predicted by sex-specific psychosocial and neurobiological features. Eur. J. Neurosci. 2019;50(3):2346–2356. doi: 10.1111/ejn.13989. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Verbruggen F., Aron A.R., Band G.P., et al. A consensus guide to capturing the ability to inhibit actions and impulsive behaviors in the stop-signal task. Elife. 2019;8 doi: 10.7554/eLife.46323. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Weigard A., Heathcote A., Matzke D., Huang-Pollock C. Cognitive modeling suggests that attentional failures drive longer stop-signal reaction time estimates in attention deficit/hyperactivity disorder. Clin. Psychol. Sci.: A J. Assoc. Psychol. Sci. 2019;7(4):856–872. doi: 10.1177/2167702619838466. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Whelan R., Conrod P.J., Poline J.B., et al. Adolescent impulsivity phenotypes characterized by distinct brain networks. Nat. Neurosci. 2012;15(6):920–925. doi: 10.1038/nn.3092. [DOI] [PubMed] [Google Scholar]
- Whelan R., Watts R., Orr C.A., et al. Neuropsychosocial profiles of current and future adolescent alcohol misusers. Nature. 2014;512(7513):185–189. doi: 10.1038/nature13402. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
ABCD data are available through the NIMH Data Archive and the code underlying the analyzes conducted in this paper are available at https://github.com/sahahn/SST_Response.