Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2021 Oct 20.
Published in final edited form as: Neurobiol Lang (Camb). 2020 Sep 1;1(3):319–338. doi: 10.1162/nol_a_00015

Brain structures and cognitive abilities important for the self-monitoring of speech errors

Ayan S Mandal (1),(2), Mackenzie E Fama (1),(3), Laura M Skipper-Kallal (1), Andrew T DeMarco (1), Elizabeth H Lacey (1),(4), Peter E Turkeltaub (1),(4)
PMCID: PMC8528269  NIHMSID: NIHMS1697592  PMID: 34676371

Abstract

The brain structures and cognitive abilities necessary for successful monitoring of one’s own speech errors remain unknown. We aimed to inform self-monitoring models by examining the neural and behavioral correlates of phonological and semantic error detection in individuals with post-stroke aphasia. First, we determined whether detection related to other abilities proposed to contribute to monitoring according to various theories, including naming ability, fluency, word-level auditory comprehension, sentence-level auditory comprehension, and executive function. Regression analyses revealed that fluency and executive scores were independent predictors of phonological error detection, while a measure of word-level comprehension related to semantic error detection. Next, we used multivariate lesion-symptom mapping to determine lesion locations associated with reduced error detection. Reduced overall error detection related to damage to a region of frontal white matter extending into dorsolateral prefrontal cortex (DLPFC). Detection of phonological errors related to damage to the same areas, but the lesion-behavior association was stronger, suggesting the localization for overall error detection was driven primarily by phonological error detection. These findings demonstrate that monitoring of different error types relies on distinct cognitive functions, and provide causal evidence for the importance of frontal white matter tracts and DLPFC for self-monitoring of speech.

Keywords: aphasia, voxel-based lesion symptom mapping, speech error detection, speech fluency, executive function, frontal white matter, dorsolateral prefrontal cortex

Introduction

Although fluent speech is littered with errors, healthy speakers can identify and repair these mistakes. Successful communication depends on this ability to self-correct. Previous studies of error detection have also shown that it predicts positive therapeutic outcomes for both production and comprehension in aphasia (R. C. Marshall, Neuburger, & Phillips, 1994). Schwartz and colleagues provided evidence that the correction of semantic errors promotes an adaptive change that allows aphasic patients to learn from their mistakes (Schwartz, Middleton, Brecher, Gagliardi, & Garvey, 2016). These results imply that self-monitoring has important consequences for aphasia recovery.

Despite the evidence that self-monitoring plays a role in language relearning, little is known about the processes underlying it. There are currently two broad categories of theories regarding self-monitoring. The first category consists of comprehension-based models, where persons detect errors by listening to their own speech (Postma, 2000). When a speaker hears himself say something different from what he intended to say, then he can identify and correct the error. Chief among comprehension-based models for error detection is the perceptual loop model initially posed by Levelt (1983). Under the perceptual loop model, an outer auditory loop monitors overt speech using speech comprehension systems. Pre-articulatory inner speech is monitored as well via an inner loop which also relies on speech comprehension (Indefrey & Levelt, 2004). This model is parsimonious because it does not assume the existence of a system dedicated to detecting errors itself. Rather, it suggests that the same mechanism that allows people to comprehend the speech of others also allows them to detect their own errors.

Comprehension-based models for self-monitoring predict that poor error detection will correlate with poor comprehension abilities. If one detects an error by comprehending one’s own overt speech, then a person who has difficulties comprehending the speech of others should also have difficulties with self-monitoring. However, Nickels and Howard found no correlation between error detection and any of three measures of auditory comprehension (1995). Furthermore, comprehension and error detection doubly dissociate: there have been case studies of patients with aphasia who accurately detect their own errors but demonstrate poor comprehension (R. C. Marshall, Rappaport, & Garcia-Bunuel, 1985) as well as patients with poor error detection yet intact comprehension (Butterworth & Howard, 1987; Liss, 1998; J. Marshall, Robson, Pring, & Chiat, 1998). It is worth noting, however, that while these studies provide substantial evidence against an overt-speech monitoring loop, it has been more difficult to test the functioning of an inner-speech monitoring loop.

In contrast to comprehension-based speech monitors, some authors have proposed production-based self-monitoring systems, where information from the speech production process itself can be used to detect an error. Particularly notable amongst production-based models for error detection is the conflict-based monitor proposed by Nozari and colleagues (2011). In Nozari’s model, the language system emits conflict signals that arise from competition between various semantic features or phonological units activated during naming. These signals are received by the anterior cingulate cortex (ACC), which processes the conflict and alerts the lateral prefrontal cortex to exert top-down control to resolve conflict (Botvinick, Braver, Barch, Carter, & Cohen, 2001; Nicholas Yeung, 2015; Nick Yeung, Botvinick, & Cohen, 2004). The conflict-based model has recently gained support from a behavioral study in neurotypical children (Hanley, Cortis, Budd, & Nozari, 2016) and a functional MRI study in healthy adults (Gauvin, De Baene, Brass, & Hartsuiker, 2015).

The self-monitoring models mentioned above make different predictions regarding the neural and behavioral correlates of error detection. The perceptual loop model predicts that detection should depend on comprehension. Consequently, the perceptual loop model also predicts detection to rely on brain structures that subserve comprehension, i.e., primarily regions within the temporal lobe (Hillis, Rorden, & Fridriksson, 2017). Production monitors in general predict that lesions to areas important for word production, broadly within the frontal lobe, should impair error detection capabilities (Catani et al., 2013; Mandelli et al., 2014). The conflict-based model in particular also predicts that regions involved in domain general cognitive processing, such as the anterior cingulate cortex (ACC) or lateral prefrontal cortex (LPFC), should also be involved (Yeung, 2015).

People with aphasia sometimes exhibit impairments in monitoring that are specific to one type of error over another, such as detecting each of their phonological errors but failing to notice any of their semantic errors (R. C. Marshall et al., 1985; Stark, 1988). Such observations suggest that different cognitive processes may be involved in the monitoring of different types of errors. This notion has received empirical support in recent years from studies that have found differences in the timing of and the learning from semantic versus phonological error monitoring (Schuchard, Middleton, & Schwartz, 2017; Schwartz et al., 2016). Comprehension-based models do not explicitly account for differential monitoring of phonological versus semantic errors, although one may expect a selective deficit in the monitoring of one type of error if the comprehension system has incurred a selective impairment in either phonological or semantic processing. However, a case study of a patient with impaired phonological auditory processing (in the form of auditory agnosia) yet preserved reading comprehension demonstrated the opposite of this expectation: the subject detected almost all of her phonological errors but ignored her semantic errors (R. C. Marshall et al., 1985). Production-based monitors for error detection have been proposed for virtually every stage of word production, from lemma selection to tactile feedback following word articulation (Postma, 2000), and therefore could support differential monitoring of errors that arise at different stages. Nozari et al. (2011) found that measures of the lexical-semantic and lexical-phonological stages of naming predicted detection of semantic and phonological errors, respectively. The authors suggested that damage to the lexical-semantic or lexical-phonologic stages of naming causes noise to predominate in the activations of representations by these systems, which obscures the ability of a monitor to detect conflict related to errors.

An examination of the brain structures and cognitive abilities associated with detection of different types of naming errors could help not only distinguish between current self-monitoring accounts but also extend existing theories into new domains.

In the present study, we tested the anatomical and behavioral predictions of the perceptual loop and conflict-based models for error detection in a group of participants with post-stroke aphasia. First, we determined whether error detection within the context of a picture naming task related to cognitive abilities proposed to contribute to monitoring, including naming ability, fluency, word-level auditory comprehension, sentence-level auditory comprehension, and executive function. Then, we used support vector regression lesion symptom mapping (SVR-LSM) to map the brain areas necessary for error detection, and probed the inter-relationships between the behavioral and neural correlates of detection ability.

Materials and Methods

Participants

Data for the current study were pooled from cohorts of left hemisphere stroke survivors recruited for two different studies at Georgetown University and MedStar National Rehabilitation Hospital. Forty-nine patients (Cohort 1) were participating in a battery of tasks to determine baseline language abilities in a transcranial direct current stimulation (tDCS) clinical trial. Fifty-four patients (Cohort 2) were participating in a study designed to probe the subjective experience of inner speech in aphasia. Twenty-three patients participated in both studies, yielding a total of 80 potential participants after combining the two cohorts. All participants were native English speakers, had no history of other brain disorder or damage, had anterior circulation strokes, and were tested at least six months after the stroke. Participants were excluded from analyses if they produced too few errors of the type being examined to assess error detection (see Dependent Variables). Demographic details of the participants included in each analysis are listed in Table 1. The study was approved by the Georgetown University Institutional Review Board, and written informed consent was obtained from all study participants prior to enrollment in the study.

Table 1. Demographic data and performance on measures of interest for study participants.

Standard deviations are presented in parentheses. Each behavioral measure has been divided by its maximum possible score, aside from MLU, digit span, and spatial span tasks which have no maximum score. Note that the Phonological Error Detection and Semantic Error Detection groups are subsets of the Total Error Detection Group. Twenty-nine participants are in both the Phonological Error Detection and Semantic Error Detection groups.

Total Error Detection Group (N=64) Phonological Error Detection Group (N=58) Semantic Error Detection Group (N=32)
Age (years) 60.4 (9.3) 60.0 (9.2) 60.6 (8.4)
Sex (M/F) 42/22 38/20 19/13
Time Since Stroke (months) 47.6 (48.8) 43.4 (44.6) 49.3 (57.2)
Education (years) 16.0 (2.9) 15.9 (3.0) 16.3 (2.6)
Handedness (L/R/AMB) 56/5/3 51/4/3 27/4/3
Lesion Size (cm3) 117.9 (81.0) 117.8 (81.4) 121.6 (80.5)
Naming Accuracy 0.52 (0.31) 0.51 (0.31) 0.48 (0.29)
Phonological Errors 0.48 (0.21) 0.51 (0.19) 0.39 (0.19)
Semantic Errors 0.16 (0.12) 0.15 (0.10) 0.24 (0.11)
Total Error Detection 0.40 (0.24) 0.40 (0.25) 0.42 (0.21)
Phonological Error Detection 0.40 (0.28) 0.39 (0.28) 0.42 (0.24)
Semantic Error Detection 0.38 (0.25) 0.38 (0.25) 0.38 (0.25)
WAB Auditory Comprehension 0.93 (0.065) 0.92 (0.065) 0.92 (0.074)
Word-to-Picture Matching 0.92 (0.11) 0.91 (0.11) 0.92 (0.10)
MLU 4.2 (2.3) 4.1 (2.3) 4.4 (2.2)
Digit Span Forwards 4.7 (3.2) 4.5 (3.2) 5.3 (3.6)
Digit Span Backwards 2.2 (2.0) 2.0 (1.9) 2.3 (2.2)
Digit Span Difference 2.5 (2.4) 2.5 (2.4) 3.0 (2.3)
Spatial Span Forwards 6.3 (2.2) 6.3 (2.2) 6.1 (2.5)
Spatial Span Backwards 5.1 (2.4) 5.3 (2.3) 5.3 (2.4)
Spatial Span Difference 1.1 (1.7) 0.95 (1.7) 0.81 (1.4)

Behavioral Tasks

All relevant tasks that had been administered to both Cohort 1 and Cohort 2 were selected for the study. These included tasks that measure confrontation picture naming performance and error detection, as well as several functions proposed to be important for error monitoring, including word-level auditory comprehension, sentence-level speech comprehension, short-term memory, working memory, executive function, and fluency. In participants for whom scores were available from both prior studies, the scores of the two equivalent tests were averaged.

Philadelphia Naming Test (PNT):

Participants from Cohort 1 were administered a 60-item version of the Philadelphia Naming Test (PNT; Roach et al., 1996). The PNT is a picture-naming task where participants must name a series of black and white drawings. Participants from Cohort 2 were administered a 120-item naming task including the 60-item PNT plus an additional 60 items. For patients who participated in both prior studies, only the 60 PNT items that matched across both groups were used to provide an overall naming accuracy score, but all available trials were pooled across both tasks for the purpose of coding error types and error detection. To determine whether this pooling substantially impacted the results, analyses were also conducted using the error detection performance from only the 60-item PNT administered to both cohorts (see Supplementary Materials).

WAB Auditory Comprehension:

The Western Aphasia Battery-Revised (WAB) Yes/No Questions was administered. This task requires a yes/no response to 20 items including questions that are biographical, environmental, and non-contextual/grammatically complex in nature (Kertesz, 2006) and assesses sentence-level comprehension ability.

Digit Span Forwards:

Participants were asked to repeat strings of numbers of increasing length in the same order in which they heard the sequence. Two strings were presented at each length, and testing stopped after both strings at a given length were recited incorrectly. The total number of strings recited correctly was taken, in which a low number indicates poor verbal short-term memory.

Word-to-Picture Matching:

Participants heard a word and pointed to the target item in a field of semantically-related pictures. The version of the task used in Cohort 1 included a field of six pictures, whereas in Cohort 2 the field included four pictures. Both tasks included 48 trials. Accuracies on the two tasks were not different in the 23 patients who performed both (Paired T(22) = .776, P = .45), so the tasks were treated as equivalent. Poor performance on this task is interpreted to indicate word-level comprehension impairment, but could reflect other factors (see Discussion).

Mean Length of Utterance (MLU):

Mean length of utterance during a picture description task was used as a measure of speech fluency. This quantity is derived by calculating the mean number of words used in each utterance. Participants from Cohort 1 described the picnic scene from the WAB, while those from Cohort 2 described the “Cookie Theft” picture from the Boston Diagnostic Aphasia Examination (Goodglass et al., 2001). The scores from the two pictures were not different in the patients for whom both scores were available (Paired T(19) = 1.20, P = .244), so the tasks were treated as equivalent. Five subjects were missing data and so are excluded from analyses using MLU.

Digit Span Backwards:

The same procedures were followed as for digit span forwards, except that participants were asked to recite the number strings in reverse order, a manipulation that is typically thought to tax executive control of working memory.

Digit Span Difference:

The difference between the forward and backward digit span scores was calculated. A high number on the measure is interpreted here to indicate poor executive control, but could also reflect other factors (see Discussion).

Spatial Span Forwards:

The Corsi-Block tapping task was used (Corsi, 1972). Participants were asked to tap a sequence of blocks of increasing length in the same order in which they saw an examiner tap the blocks. Two sequences were presented at each length, and testing stopped after both sequences of a given length were repeated incorrectly. The total number of sequences repeated correctly was taken, in which a low number indicates poor nonverbal short-term memory.

Spatial Span Backwards:

The same procedures were followed as for spatial span forwards, except that participants were asked to tap the sequence of blocks in reverse order, a manipulation that is typically thought to tax executive control of working memory.

Spatial Span Difference:

The difference between the forward and backward spatial span scores was calculated. A high number on this measure is interpreted to indicate poor executive control, but could also reflect other factors (see Discussion).

Coding Naming Responses for Error Type and Error Detection

Error Type:

Videos of naming responses were transcribed into the International Phonetic Alphabet and scored offline for accuracy, error type, and error detection. Error coding was based on the PNT rules (Roach et al., 1996). Only the first naming attempt for each item was coded. False starts and fragments were not considered as first naming attempts as per PNT scoring rules. Errors, whether words or nonwords, were coded as phonological if they shared the stressed vowel, at least two phonemes, or the first or last phoneme as the target. Errors that were semantically related to the target were coded as semantic errors. Errors that were both phonologically and semantically related to the target were coded with mixed errors and were thus considered as neither phonological nor semantic. The other error types specified in the PNT scoring rules were coded, but are not considered here. Seventeen subjects (four of whom received the 120-item naming task) were graded by two independent scorers to determine interrater reliability. Errors received the same error code from different scorers on 91.7% of trials. This is comparable to the interrater reliability observed in similar studies (Schwartz et al., 2016).

Error Detection:

Error detection was assessed within the context of the picture naming task using an adapted version of the coding protocol developed by Schwartz and colleagues (Schwartz et al., 2016). A participant was said to have detected an error if he/she uttered a statement indicating awareness that an error was made or if he/she attempted to self-correct the error. In other words, when a participant made an error, it would count as detected if the person either followed the errant response with a statement like “no, not that”, or made an attempt to correct their first response (e.g., “cat… I mean dog”), indicating recognition that the first response was incorrect. Attempts at self-correction were coded as detected errors regardless of whether the second attempt was correct or incorrect because either provides evidence that the individual is aware of the error. Only the first naming attempt for each item was scored for error detection; detection or correction of errors in attempts at self-correction were not scored (e.g., for “cat… I mean dog… I mean bird” only the first self-correction is scored). This ensured that there was only one error detection score per item, which reduced bias toward individuals who made many attempts at self-correction and makes it simpler to interpret detection of specific error types. Nonverbal indicators of error acknowledgement, such as head shakes, were not coded. Repetition of the initial naming attempt (e.g., “cat…cat”) was also not coded as an error detection. The accuracy of attempts at self-correction was recorded, but those data are not considered here because there was not adequate power to further divide the types of detections and corrections. Consistent with protocols in prior studies (Nozari et al., 2011; Schuchard et al., 2017; Schwartz et al., 2016), participants were not given instructions to indicate awareness of an error or to self-correct, so all error detection was spontaneous.

Dependent Variables

Total Error Detection:

The total error detection score was calculated by dividing the number of detections by the total number of errors on initial naming attempts, and so is expressed as the proportion of errors that were detected by each individual. Errors of all types – phonological, semantic, mixed, etc. – were considered for this measure. Total error detection scores were not considered from participants who produced fewer than ten errors, leaving 64 participants in the analyses of total error detection.

Phonological Error Detection:

The phonological error detection score was calculated by dividing the number of detections of phonological errors by the total number of phonological errors committed. Phonological error detection scores were not considered from participants who produced fewer than five phonological errors, leaving 58 participants in the analyses of phonological error detection.

Semantic Error Detection:

The semantic error detection score was calculated by dividing the number of detections of semantic errors by the total number of semantic errors committed. Semantic error detection scores were not considered from participants who produced fewer than five semantic errors, leaving 32 participants in the analyses of semantic error detection.

Behavioral Analysis

Statistical analyses were conducted in SPSS 25 (https://www.ibm.com/products/spss-statistics). Outlier scores (greater or less than three inter-quartile distances from the median) were excluded from analyses. This resulted in exclusion of two individuals’ word-picture matching scores.

A series of one-way ANOVAs was first used to confirm that scores of each test did not differ between the participants from only Cohort 1, only Cohort 2, and those who were in both cohorts.

Bivariate correlations were used to screen for inter-relationships between variables of interest. Next, three multiple linear regression analyses were performed with backward elimination of predictors with P > .1 to determine which behavioral scores related to detection of each error type (all errors, phonological errors, semantic errors). Only scores which were univariately correlated (uncorrected for multiple comparisons) with at least one of the dependent variables were entered into these regression analyses. These scores included word-to-picture matching, MLU, digit span difference, spatial span backwards, and spatial span difference.

To investigate whether the monitoring of different types of errors related to distinct cognitive abilities, we performed two additional regression analyses for phonological and semantic error detection, where detection of the other error type was included as predictor along with the original cognitive measures. These analyses were necessarily limited to individuals who were in both the phonological error detection and semantic error detection groups and had scores for all behavioral predictors (N= 27).

Neuroimaging

MRI Acquisition and Pre-Processing

3D T1-weighted MRIs were acquired from participants on a 3.0 T Siemens Trio scanner with the following parameters: TR = 1900 ms; TE = 2.56 ms; flip angle = 9°; 160 contiguous 1 mm sagittal slices; field of view (FOV) = 250 × 250 mm; matrix size = 246 × 256, voxel size 1 mm3. A T2-weighted sampling perfection with application optimized contrasts using different flip angle evolution (SPACE) sequence was acquired with the following parameters: 176 sagittal slices; slice thickness = 1.25 mm, FOV = 240 240 mm; matrix size = 384 384; TR = 3200 ms; echo train length = 145, variable TE; variable flip angle, voxel size = 0.625 0.625 1.25 mm.

Lesions were manually traced on co-registered T1-weighted and T2-weighted images in native space using ITK-SNAP 3.6 (http://www.itksnap.org) by a board certified neurologist (P.E.T.). Native space MPRAGEs and lesion tracings were warped to MNI space using the Clinical Toolbox Older Adult Template as the target template (Rorden, Bonilha, Fridriksson, Bender, & Karnath, 2012) via a custom pipeline. First, brain parenchyma was extracted from each native space image by applying a mask intended to minimize the clipping of gray matter edges. The initial mask was generated by combining the lesion tracing image (binarized) with white and gray matter tissue probability maps generated by the unified segmentation procedure in SPM12 (https://www.fil.ion.ucl.ac.uk/spm/software/spm12/) applied to the original native space image, cost-function masked with the lesion tracing. The resulting mask was blurred and inverted to remove non-brain tissue from the image. The resulting brain extracted image was then normalized using Advanced Normalization Tools software (ANTs; http://picsl.upenn.edu/software/ants/; (Avants, Tustison, & Song, 2009)). Lesion masking was used at each step of the ANTs process. After bias field correction was applied, normalization proceeded using a typical ANTs procedure, including a rigid transform step, an affine transform step, and a nonlinear SyN step. Next, the output of this initial ANTs warp was recursively submitted to three additional applications of the SyN step. Finally, the resulting linear (rigid and affine) and four nonlinear warp fields were concatenated and the original native space MPRAGE and lesion tracings were transformed to the template space using BSpline interpolation. This iterative application of nonlinear warping was intended to improve normalization of expanded ventricles and displaced deep structures in individuals with large lesions. The normalized lesion tracings were finally downsampled to 2.5 mm3.

Lesion-Symptom Mapping

We implemented SVR-LSM (Zhang, Kimberg, Coslett, Schwartz, & Wang, 2014) using a MATLAB-based toolbox (DeMarco and Turkeltaub, 2018) running under MATLAB R2017a (The MathWorks, Inc., Natick, Massachusetts, United States). SVR-LSM was used to identify the left-hemisphere areas associated with impaired self-monitoring. SVR-LSM applies a machine learning based algorithm to find lesion-symptom relationships more sensitively and specifically than traditional mass-univariate lesion-symptom mapping approaches (Mah, Husain, Rees, & Nachev, 2014). Only voxels damaged in at least 10% of the participants in the study were considered for each analysis. Lesion volume confounds were controlled in all analyses by regressing the lesion volume out of both behavioral scores and lesion masks, a method that provides rigorous control of lesion volume and is more sensitive than alternative approaches (DeMarco and Turkeltaub, 2018). Voxelwise beta-values were thresholded at p < .005 using 10,000 permutations of the behavioral scores to generate voxelwise null distributions. To correct for multiple comparisons, a cluster threshold determined from the 10,000 permutation maps was applied to control the family wise error rate at .05 (Mirman et al., 2018). SVR-LSM analyses were performed examining lesion locations associated with failure to detect each error type (total errors, phonological errors, semantic errors). An additional SVR-LSM analysis for phonological error detection was performed where mean length of utterance (MLU) was added as a covariate. MRIs were not obtained on some participants, so the sample sizes for the SVR-LSM analyses were N=57 for total errors, N=51 for phonological errors, and N=29 for semantic errors.

Model quality was assessed in two ways. The first was prediction accuracy, which is a density of correlation coefficients between predicted scores and training scores across 10 replications of a 5-fold cross-validated model. The mean of this density (average correlation coefficient) is used to summarize how well the predicted scores trend with the real scores. However, previous work has observed that the quality of back-projected spatial patterns cannot be assessed on the basis of prediction accuracy alone (Rasmussen et al., 2012). Indeed, this work has observed a trade-off between model visualization reproducibility and prediction accuracy. Therefore, a second metric produced by Rasmussen et al. (2012), pattern reproducibility index, was used to assess reproducibility of the back-projected pattern. Pattern reproducibility index is calculated as a density of voxelwise correlation coefficients computed pairwise between 10 replicates of SVR-β maps, each generated using a random 80% of observations. Model quality measures were as follows: Total Detection accuracy 0.31 (SD 0.07), reproducibility r = 0.88 (SD 0.05); Phonological Detection accuracy 0.25 (SD 0.07), reproducibility r = 0.89 (SD 0.04); Phonological Detection controlling MLU accuracy 0.26 (SD 0.11), reproducibility r = 0.86 (SD 0.04); Semantic Detection accuracy 0.36 (SD 0.15), reproducibility r = 0.75 (SD 0.12).

Region of interest (ROI) Analyses for Semantic Error Detection

To compensate for the smaller sample size of the semantic error detection SVR-LSM analyses, we selected theory-driven regions of interest (ROIs) from the Harvard Oxford Cortical atlas to determine whether brain regions important for auditory comprehension or speech production related to semantic error monitoring. Selected ROI’s included the posterior superior temporal gyrus (pSTG), anterior superior temporal gyrus (aSTG), posterior middle temporal gyrus (pMTG), Heschl’s gyrus (HG), planum temporale (PT), angular gyrus (AG), inferior frontal gyrus, pars triangularis (IFG), and middle frontal gyrus (MFG). Lesion load for each ROI was calculated for each participant by dividing the number of voxels which overlapped between the lesion mask and ROI by the total number of voxels within the ROI. Partial correlations were calculated between ROI lesion load and semantic error detection, controlling for total lesion volume.

Data Availability

The data that support the findings of this study are available on request from the corresponding author. The data are not publicly available due to their containing information that could compromise the privacy of research participants. Source code that was used to conduct the SVR-LSM analyses in this study is available at https://github.com/atdemarco/svrlsmgui/.

Results

Relationships of Behavioral Scores to Error Detection

Average scores across the groups are shown in Table 1. Since the data were derived from two partially overlapping patient cohorts, we first performed a series of one-way ANOVAs to confirm that cohort differences did not introduce biases in the variables of interest (age, time since stroke, education, PNT, word-to-picture matching, WAB Auditory Comprehension, MLU, digit span forwards, digit span backwards, digit span difference, spatial span forwards, spatial span backwards, spatial span difference, total error detection, phonological error detection, semantic error detection). No effects of cohort (participants from only cohort 1 vs. participants from only cohort 2 vs. individuals from both cohorts) were identified (all P > .05).

Bivariate correlations between behavioral variables of interest and each dependent variable are provided in Table 2 and Table 3. Intercorrelations among behavioral variables of interest are provided in Table 4. Phonological error detection and semantic error detection were correlated with one another (r = .570; p = .0012; N=29), and this relationship persisted after controlling for lesion volume (r = .502; p = .011; N=23). Regression models were used to identify predictors of error detection, first examining detection of all errors, and then semantic and phonological errors separately. Only behavioral measures which were univariately correlated with at least one of the dependent variables were entered into these regression models. These measures were word-to-picture matching, MLU, digit span difference, spatial span backwards, and spatial span difference. Total error detection was predicted by MLU (standardized β = .380; p = .002; VIF = 1.025) and spatial span difference (standardized β = −.333; p =.006; VIF = 1.025) with an overall Adjusted R2 of .268 (F(2,54) = 11.3, p < .001). Phonological error detection was predicted by MLU (standardized β = .428; p =.001; VIF = 1.031), digit span difference (standardized β = −.249; p =.04; VIF = 1.061), and spatial span difference (standardized β = −.255; p =.038; VIF = 1.091) with an overall Adjusted R2 of .334 (F(3,48) = 9.54; p < .001). Semantic error detection was predicted only by word-to-picture matching (standardized β = .516; p = .003; VIF = 1.000) with an overall Adjusted R2 of .240 (F(1,28)=10.173, p = .003). These results were mostly preserved when reanalyzed using error detection performance based only on the 60 PNT items administered to all participants (see Supplementary Material).

Table 2.

Correlations between error detection and demographic variables.

Total Error Detection (N=64) Phonological Error Detection (N=58) Semantic Error Detection (N=32)
Age .269* .200 −.183
Time Since Stroke −.035 −.044 .080
Education .211 .101 −.001
***

P<.05 after Bonferroni correction for 9 comparisons,

**

P<.01 uncorrected,

*

P<.05 uncorrected.

Table 3. Correlations between error detection and other variables.

Sample sizes differ due to missing scores.

Total Error Detection Phonological Error Detection Semantic Error Detection
PNT .027 (N=64) −.063 (N=58) .079 (N=32)
WAB Auditory Comprehension .115 (N=62) .022 (N=56) .139 (N=32)
Word-to-Picture Matching .302* (N=62) .244 (N=56) .523** (N=31)
MLU .514*** (N=59) .502*** (N=54) .365 (N=31)
Digit Span Forwards −.074 (N=59) −.210 (N=54) .033 (N=31)
Digit Span Backwards .148 (N=59) .084 (N=54) .082 (N=31)
Digit Span Difference −.226 (N=59) −.340** (N=54) −.023 (N=31)
Spatial Span Forwards .096 (N=59) .022 (N=54) .265 (N=31)
Spatial Span Backwards .327** (N=59) .276* (N=54) .377* (N=31)
Spatial Span Difference −.329** (N=59) −.352* (N=54) −.171 (N=31)
***

P<.05 after Bonferroni correction for 33 comparisons,

**

P<.01 uncorrected,

*

P<.05 uncorrected.

Table 4. Correlations between behavioral measures.

Correlations are performed for the full sample (N=64). Sample size varies between 57 and 64 on individual correlations because of missing values on some tests.

WAB Yes-No Word-to-Picture Matching MLU Digit Span Fwd Digit Span Bwd Digit Span Diff Spatial Span Fwd Spatial Span Bwd Spatial Span Diff
PNT .277* .679*** .491*** .562*** .529*** .313* .314* .253* .055
WAB Yes-No .337** .332* .471*** .304* .374** .140 .032 .137
Word-to-Picture Matching .456*** .340** .440*** .079 .401*** .416*** −.069
MLU .474*** .626*** .092 .190 .252 −.104
Digit Span Fwd .680*** .780*** .379** .261* .128
Digit Span Bwd .071 .474*** .490*** −.065
Digit Span Diff .110 −.063 .230
Spatial Span Fwd .723*** .292*
Spatial Span Bwd −.450***
***

P<.05 after Bonferroni correction for 33 comparisons,

**

P<.01 uncorrected,

*

P<.05 uncorrected.

To determine whether the behavioral predictors of phonological and semantic error detection were specific to the monitoring of those respective error types, we performed two additional regression analyses for phonological and semantic error detection, where detection of the other error type was included as a predictor along with the original cognitive measures. Phonological error detection was predicted by semantic error detection (standardized β = .587; p = .001; VIF = 1.23), spatial span difference (standardized β = −.321; p = .039; VIF = 1.02), spatial span backwards (standardized β = −.336; p = .047; VIF = 1.23), and MLU (standardized β = .283; p = .071; VIF = 1.08) with an overall Adjusted R2 of .461 (F(4,22) = 6.56; p = .001). Semantic error detection was predicted phonological error detection (standardized β = .392; p = .018; VIF = 1.14), and word-to-picture matching (standardized β = .464; p = .006; VIF = 1.14) with an overall Adjusted R2 of .455 (F(2,24) = 11.8; p < .001).

Lesion-symptom mapping

Lesion overlap maps demonstrated good coverage of the left middle cerebral artery territory for all analyses (Figure 1). Lesion volume correlated modestly with total error detection (r = −.278; p = .04), and a trending relationship was observed with phonological error detection (r = −.243; p = .09) and semantic error detection (r = −.31; p = .10). Lesion volume was subsequently controlled for in all SVR-LSM analyses. SVR-LSM analysis demonstrated that decreased total error detection mapped onto lesions in a large region of frontal white matter and DLPFC (Figure 2). A single cluster was identified with a volume of 7156 mm3, with center of mass at MNI coordinates −32.7, 8.7, 20.5. Decreased detection of phonological errors related to damage to the same regions as total error detection, but the relationship was stronger, resulting in a larger cluster of significant lesion-deficit association (8046 mm3), centered at −31.8, 9.4, 26.1 (Figure 3). The stronger lesion-behavior relationship observed for phonological error detection suggests that the localization of total error detection was driven primarily by phonological error detection. To confirm that the results did not simply reflect reduced speech output among individuals with frontal white matter damage, we added MLU as a covariate to our SVR-LSM analysis of phonological error detection. The result remained significant after controlling for fluency and centered more closely to DLPFC. A single cluster was identified with a volume of 9641 mm3, and center of mass at MNI coordinates −37.3, 9.3, 28.7 (Figure 4). No significant clusters were identified in the SVR-LSM analysis of semantic error detection (Supplementary Figure 1). These results did not change substantially when reanalyzed using error detection performance based only on the 60 PNT items administered to all participants (see Supplementary Material).

Figure 1. Lesion overlap maps for study participants.

Figure 1.

Lesion symptom mapping analyses were limited to voxels that were lesioned in at least 10% of participants (i.e. at least 6 participants for total and phonological error detection, at least 3 for semantic error detection).

Figure 2. Multivariate lesion symptom mapping results for total error detection.

Figure 2.

SVR-LSM analysis demonstrated that decreased detection of all error types was associated with damage to a large region of frontal white matter and dorsolateral prefrontal cortex (voxelwise P<.005, cluster-level familywise error <.05).

Figure 3. Multivariate lesion symptom mapping results for phonological error detection.

Figure 3.

Reduced detection of phonological errors related to damage to frontal white matter and dorsolateral prefrontal cortex (voxelwise P<.005, cluster-level familywise error <.05).

Figure 4. Multivariate lesion symptom mapping results for phonological error detection, controlling for fluency.

Figure 4.

After adding MLU as a covariate, reduced detection of phonological errors related to damage to dorsolateral prefrontal cortex (voxelwise P<.005, cluster-level familywise error <.05).

ROI Analyses for Semantic Error Detection

In light of the reduced sample size for the SVR-LSM analysis of semantic error detection, theory-driven ROIs were selected to determine the association of semantic error detection with lesions in specific brain regions involved in auditory comprehension, lexical processing, word production, and executive function (Table 5). The results of all ROI analyses were null, even at an uncorrected statistical threshold.

Table 5. Results of ROI analyses.

Partial correlations between ROI lesion load and semantic error detection, controlling for total lesion volume. Negative values indicate a relationship between ROI damage and behavioral impairment.

Semantic Error Detection
(N=29)
aSTG .035
pSTG .298
pMTG .153
Heschl’s gyrus .126
Planum Temporale .208
Angular Gyrus .297
IFG −.065
MFG −.167
***

P<.05 after Bonferroni correction for 8 comparisons,

**

P<.01 uncorrected,

*

P<.05 uncorrected.

Discussion

In this study, we examined the behavioral and lesion correlates of impaired detection of speech errors in people with aphasia. We found that reduced detection of phonological errors was associated with low fluency and low performance on measures of executive control. Reduced semantic error detection related to low performance on a test of word-level comprehension ability. Finally, we found that damage to frontal white matter and the DLPFC was associated with poor error detection. Overall, these findings provide valuable evidence regarding the neural and behavioral underpinnings of speech error monitoring, and suggest that deficits in monitoring of phonological and semantic errors relate to somewhat different mental processes. Findings are consistent with aspects of both production- and comprehension-based models of error monitoring.

Support for production-based error monitoring models

Fluency marks successful speech production, and in this sense, an association between fluency and phonological error detection supports production-based monitors in general. A more precise account of this relationship is less clear. The Nozari conflict model predicts phonological error monitoring to decrease as a result of higher noise in the system responsible for the activation of phonological nodes (Nozari et al., 2011). Low fluency could reflect increased noise in the phonological system causing delays in retrieval, but this interpretation is not consistent with our lesion-symptom mapping result, given that phonological error production has been associated with lesions of the parietal lobe (Mirman et al., 2015; Schwartz, Faseyitan, Kim, & Coslett, 2012). Alternatively, since both fluency and phonological errors are associated with deficits in post-lexical speech production (Nickels & Howard, 1995; Romani, Olson, Semenza, & Granà, 2002; Schwartz, Wilshire, Gagnon, & Polansky, 2004; Wilshire, 2002), the association between fluency and detection of phonological errors may imply that these errors are detected during post-lexical speech production. Although the Nozari conflict model does not include post-lexical speech production stages, many production-based models of error monitoring suggest that monitoring also occurs at these stages (see Postma, 2000 for a review).

Our lesion symptom mapping results are also consistent with models suggesting that error detection relies on speech production processes. Reduced error detection related to damage to a large region of frontal white matter underlying areas associated with speech production (Price, 2010). Prior studies have indicated that white matter tracts in this region are responsible for supporting fluent speech (Catani et al., 2013; Mandelli et al., 2014), which is consistent with our behavioral findings. A recent fMRI study in healthy participants also provided support for a speech monitor dependent on frontal regions. Gauvin and colleagues found that self-monitoring elicits activity across several areas involved in speech production, such as pre-SMA and IFG (Gauvin et al., 2015). Our results add causal evidence that frontal white matter tracts are involved in successful detection of speech errors.

Recent work has implicated the left frontal aslant tract (FAT), which connects left IFG to pre-SMA and SMA, in the initiating, sequencing, and stopping of language production (Catani et al., 2013; Dick, Garic, Graziano, & Tremblay, 2018). Given that the white matter implicated in our study was adjacent to IFG, it is logical to ask whether the FAT is involved in self-monitoring. However, a number of white matter tracts run through the areas implicated here, including tracts connecting medial and lateral frontal regions such as the FAT, tracts connecting subcortical structures and frontal regions, and tracts connecting frontal regions with temporoparietal areas. A future diffusion tensor imaging study could clarify the specific white matter connections that are essential for proper speech error monitoring.

The role of executive control in speech error monitoring

Considerable debate centers on whether the monitor used to detect one’s own speech errors is domain-general or language-specific. The perceptual loop model predicts the monitor to be specific to language since it should depend on the speech comprehension system (Levelt, 1983; Postma, 2000). Nozari’s conflict-based model is a domain general model in that conflict signals arising during naming are received by a frontal brain structure (namely, the ACC) which monitors errors produced in other cognitive domains as well (Nozari et al., 2011). Domain-general theories of error monitoring predict successful detection to rely on executive functioning, regardless of the type of error being detected (Botvinick et al., 2001; Yeung et al., 2004). In support of this prediction, we identified a relationship between two measures of executive function (digit span difference and spatial span difference) and the monitoring of phonological errors. However, in contrast to the predictions of a domain-general theory, we did not find an association between these measures and semantic error detection. The finding that the monitoring of different types of errors relates to different behavioral measures may be interpreted as evidence against a domain-general model. However, it is also worth noting that we found a correlation between phonological error detection and semantic error detection, implying that some overlapping processes underlie monitoring of both error types. We also found some support for the anatomical predictions of a domain-general model for error monitoring, at least in the analysis of phonological errors. Reduced error detection related to damage to the DLPFC, a region known to be associated with working memory and domain general executive function (Miller, 2001; Miller & Cohen, 2001).

Nozari’s conflict-based model predicts the anterior cingulate cortex to be involved in error monitoring, and prior evidence in healthy participants indicates that internal speech monitoring recruits ACC activity (Gauvin et al., 2015). Frontal white matter lesions may disrupt connectivity to the ACC, resulting in impaired error detection (Hogan, Vargha-Khadem, Saunders, Kirkham, & Baldeweg, 2006), but since we lacked lesion coverage of the ACC in our participant group, we cannot provide evidence for or against a causal role for this region in speech error monitoring. Future work using connectome-based lesion methods that can detect disconnected regions outside the lesioned area (Gleichgerrcht, Fridriksson, Rorden, & Bonilha, 2017; Yourganov, Fridriksson, Rorden, Gleichgerrcht, & Bonilha, 2016), will be useful to examine this issue further.

It is worth considering whether our tasks for executive control, difference between forward and backward digit/spatial span, are appropriate measures of the type of control required to resolve conflict within the framework of conflict monitoring theory. Within the conflict monitoring framework, top-down control is required to override a habitual response (e.g. such as the habitual response to read the word instead of name the color in the Stroop task) and is provided by the DLPFC (MacDonald, Cohen, Andrew Stenger, & Carter, 2000; Yeung, 2015). This type of control is likely at play when completing the backward digit or spatial span task, since the habitual response of repeating a sequence in the same order as it was presented must be overridden and replaced with a manipulation of that sequence in the reverse direction. Subtracting the forward digit/spatial span score from the backward score adjusts for the demands on maintenance in short term memory required for both tasks, isolating the executive control component in the resulting measure. However, other factors could also contribute to a difference in performance on the forward versus backward digit/spatial span tasks. Some evidence indicates that separate mechanisms, one based on refreshing and another based on rehearsal, are responsible for maintaining short term memory during low demand tasks like forward digit span on the one hand and high demand tasks like backwards digit span on the other (Baddeley, 2003; V. Camos, Lagner, & Barrouillet, 2009; Valérie Camos, Mora, & Oberauer, 2011; Ghaleh et al., 2019). Therefore, a selective deficit in backwards but not forwards digit span could reflect an impairment in systems that support the articulatory rehearsal of phonological information (i.e. the phonological loop) (Ghaleh et al., 2019), and it could be this aspect of the digit span difference score that is associated with poor error detection. The finding that the difference between backwards and forwards spatial span also related to phonological error detection makes this interpretation less likely, assuming that rehearsal for backwards spatial span relies on a visuospatial sketchpad rather than the phonological loop. Another area of less clarity is the null relationship between semantic error detection and measures of executive function. The null finding is inconsistent with a domain general model, but it may simply reflect reduced power for the semantic error detection analyses. Alternatively, our behavioral measures could be less sensitive to the type of control necessary for manipulation of semantic information. Future work comparing error detection ability with performance on a wide range of executive function and working memory tasks could clarify which cognitive abilities are required to successfully monitor speech errors.

Mechanisms of semantic error detection

Our finding that the detection of semantic errors relies on word-level comprehension is consistent with comprehension-based models of monitoring that suggest that monitoring impairments follow an inability to comprehend one’s own speech (Levelt, 1983). While prior studies did not find a relationship between auditory comprehension and error monitoring (J. Marshall et al., 1998; R. C. Marshall et al., 1985; Nickels & Howard, 1995), those studies investigated phonological, not semantic errors. A role for comprehension in semantic error monitoring would account for evidence of poor semantic error detection among patients with severe comprehension deficits (R. C. Marshall et al., 1985).

However, some evidence suggests that both word comprehension and word production rely on the same lexical-semantic representations (Foygel, 2000; Hanley & Nickels, 2009). An alternative interpretation of the word-to-picture matching measure is as an indicator of lexical-semantic access ability, which is required for naming as well as comprehension (Hanley & Nickels, 2009). Consistent with this interpretation, word-picture matching scores and naming (PNT) scores correlated more strongly with each other than either did with any other measure examined in our participants (Table 4). Semantic errors occur primarily during the lexical-semantic stage of word production due to a failure to activate the correct lemma from semantics (Foygel, 2000; Schwartz et al., 2009). As noted above, according to the conflict-based model, damage to lexical-semantic access is expected to result in high levels of conflict that cause difficulty detecting semantic errors (Nozari et al., 2011). Thus, the relationship between word-picture matching and semantic error detection could be viewed as consistent with this model.

While lesion-behavior associations in the SVR-LSM or ROI analyses might have helped clarify the mechanisms by which semantic errors are monitored, the null findings of these analyses proved unhelpful. These analyses were limited by the small sample size available. Additional studies with larger populations and more specific behavioral measures will be needed to identify the neural and behavioral correlates of semantic error detection.

Limitations

In order to capture the reflexive aspect of natural error detection, we did not give participants explicit instructions to comment on the accuracy of their naming attempts, consistent with past studies of self-monitoring in people with aphasia (Nozari et al., 2011; Schuchard et al., 2017; Schwartz et al., 2016). However, it is possible that some patients were aware of the errors they were committing, but chose not to comment on them. While in theory this issue could be addressed using a paradigm in which participants are required to report their accuracy following each naming trial, this has the potential to alter the way they approach error monitoring. We also note that combining the two cohorts of patients was necessary to achieve an adequate sample size, and behavioral measures examined here were selected as the best measures available in both cohorts. Additional measures, for instance examining non-verbal semantics, speech perception, or motor speech production, would have been informative but were not available in both cohorts. Future prospective studies of error monitoring should select measures that allow a more precise and comprehensive delineation of the behavioral correlates of error detection.

Conclusion

Gaining a more detailed understanding of the brain and behavioral basis of speech self-monitoring may lead to new treatments aimed at improving awareness of errors and self-correction in aphasia. These results demonstrate that monitoring of different error types relies on distinct cognitive functions, and provide causal evidence for the importance of frontal white matter tracts and DLPFC for self-monitoring of speech. These findings substantially inform the debate regarding the neural and behavioral underpinnings of speech error monitoring.

Supplementary Material

Supplemental Materials

Acknowledgements:

We thank Zainab Anbari, Maryam Ghaleh, Katherine Spiegel, Mary Henderson, and Harshini Pyata for contributing to data collection, and our participants for their involvement in the study.

Funding Sources:

This study was supported by the NIH/NCATS via the Georgetown-Howard Universities Center for Clinical and Translational Science (KL2TR000102), The Doris Duke Charitable Foundation (2012062), NIH/NIDCD (R03DC014310), NIH/NIDCD (F31DC014875), NIH/NCATS (TL1TR001431), NIH/NIDCD (R01DC014960).

Footnotes

Conflict of interest: The authors report no competing financial interests.

References

  1. Avants BB, Tustison N, & Song G (2009). Advanced Normalization Tools (ANTS). Insight Journal, 1–35. [Google Scholar]
  2. Baddeley A (2003). Working memory: Looking back and looking forward. Nature Reviews Neuroscience. [DOI] [PubMed] [Google Scholar]
  3. Botvinick MM, Braver TS, Barch DM, Carter CS, & Cohen JD (2001). Conflict Monitoring and Cognitive Control. Psychological Review, 108(3), 624–652. [DOI] [PubMed] [Google Scholar]
  4. Butterworth B, & Howard D (1987). Paragrammatisms. Cognition, 26(1), 1–37. [DOI] [PubMed] [Google Scholar]
  5. Camos V, Lagner P, & Barrouillet P (2009). Two maintenance mechanisms of verbal information in working memory. Journal of Memory and Language. [Google Scholar]
  6. Camos Valérie, Mora G, & Oberauer K (2011). Adaptive choice between articulatory rehearsal and attentional refreshing in verbal working memory. Memory and Cognition. [DOI] [PubMed] [Google Scholar]
  7. Catani M, Mesulam MM, Jakobsen E, Malik F, Martersteck A, Wieneke C, … Rogalski E (2013). A novel frontal pathway underlies verbal fluency in primary progressive aphasia. Brain, 136(8), 2619–2628. [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Corsi PM (1972). Human memory and the medial temporal region of the brain. Dissertation Abstract International. [Google Scholar]
  9. Demarco AT, & Turkeltaub PE (2018). A multivariate lesion symptom mapping toolbox and examination of lesion-volume biases and correction methods in lesion-symptom mapping. Human Brain Mapping. [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Dick AS, Garic D, Graziano P, & Tremblay P (2018). The frontal aslant tract (FAT) and its role in speech, language and executive function. Cortex. [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Foygel D (2000). Models of Impaired Lexical Access in Speech Production. Journal of Memory and Language, 43(2), 182–216. [Google Scholar]
  12. Gauvin HS, De Baene W, Brass M, & Hartsuiker RJ (2015). Conflict monitoring in speech processing: An fMRI study of error detection in speech production and perception. NeuroImage, 126, 96–105. [DOI] [PubMed] [Google Scholar]
  13. Ghaleh M, Lacey EH, Fama ME, Anbari Z, DeMarco AT, & Turkeltaub PE (2019). Dissociable Mechanisms of Verbal Working Memory Revealed through Multivariate Lesion Mapping. Cerebral Cortex. [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Gleichgerrcht E, Fridriksson J, Rorden C, & Bonilha L (2017). Connectome-based lesion-symptom mapping (CLSM): A novel approach to map neurological function. NeuroImage: Clinical. [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Goodglass Harold; Kaplan Edith; Barresi B. (2001). The Assessment of Aphasia and Related Disorders. Lippincott Williams & Wilkins. [Google Scholar]
  16. Hanley JR, Cortis C, Budd MJ, & Nozari N (2016). Did I say dog or cat? A study of semantic error detection and correction in children. Journal of Experimental Child Psychology, 142, 36–47. [DOI] [PubMed] [Google Scholar]
  17. Hanley JR, & Nickels L (2009). Are the same phoneme and lexical layers used in speech production and comprehension? A case-series test of Foygel and Dell’s (2000) model of aphasic speech production. Cortex, 45(6), 784–790. [DOI] [PubMed] [Google Scholar]
  18. Hillis AE, Rorden C, & Fridriksson J (2017). Brain regions essential for word comprehension: Drawing inferences from patients. Annals of Neurology, 81(6), 759–768. [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Hogan AM, Vargha-Khadem F, Saunders DE, Kirkham FJ, & Baldeweg T (2006). Impact of frontal white matter lesions on performance monitoring: ERP evidence for cortical disconnection. Brain, 129(8), 2177–2188. [DOI] [PubMed] [Google Scholar]
  20. Indefrey P, & Levelt WJM (2004). The spatial and temporal signatures of word production components. Cognition, 92(1–2), 101–144. [DOI] [PubMed] [Google Scholar]
  21. Kertesz A (2006). The Western aphasia battry -- revised. Grune & Station. New York. [Google Scholar]
  22. Levelt WJM (1983). Monitoring and self-repair in speech. Cognition, 14(1), 41–104. [DOI] [PubMed] [Google Scholar]
  23. Liss JM (1998). Error-revision in the spontaneous speech of apraxic speakers. Brain and Language, 62(3), 342–360. [DOI] [PubMed] [Google Scholar]
  24. MacDonald AW, Cohen JD, Andrew Stenger V, & Carter CS (2000). Dissociating the role of the dorsolateral prefrontal and anterior cingulate cortex in cognitive control. Science. [DOI] [PubMed] [Google Scholar]
  25. Mah YH, Husain M, Rees G, & Nachev P (2014). Human brain lesion-deficit inference remapped. Brain, 137(9), 2522–2531. [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. Mandelli ML, Caverzasi E, Binney RJ, Henry ML, Lobach I, Block N, … Gorno-Tempini ML (2014). Frontal white matter tracts sustaining speech production in primary progressive aphasia. J Neurosci, 34(29), 9754–9767. [DOI] [PMC free article] [PubMed] [Google Scholar]
  27. Marshall J, Robson J, Pring T, & Chiat S (1998). Why does monitoring fail in jargon aphasia? comprehension, judgment, and therapy evidence. Brain and Language, 63(1), 79–107. [DOI] [PubMed] [Google Scholar]
  28. Marshall RC, Neuburger SI, & Phillips DS (1994). Verbal self-correction and improvement in treated aphasic clients. Aphasiology, 8(6), 535–547. [Google Scholar]
  29. Marshall RC, Rappaport BZ, & Garcia-Bunuel L (1985). Self-monitoring behavior in a case of severe auditory agnosia with aphasia. Brain and Language, 24(2), 297–313. [DOI] [PubMed] [Google Scholar]
  30. Miller EK (2001). The prefrontal cortex and cognitive control. Nature Reviews Neuroscience, 1(1), 59–65. [DOI] [PubMed] [Google Scholar]
  31. Miller EK, & Cohen JD (2001). An integrative theory of prefrontal cortex function. Annual Review of Neuroscience, 24(1), 167–202. [DOI] [PubMed] [Google Scholar]
  32. Mirman D, Chen Q, Zhang Y, Wang Z, Faseyitan OK, Coslett HB, & Schwartz MF (2015). Neural organization of spoken language revealed by lesion–symptom mapping. Nature Communications, 6, 6762. [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Mirman D, Landrigan JF, Kokolis S, Verillo S, Ferrara C, & Pustina D (2018). Corrections for multiple comparisons in voxel-based lesion-symptom mapping. Neuropsychologia, 115, 112–123. [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Nickels L, & Howard D (1995). Phonological errors in aphasic naming: comprehension, monitoring and lexicality. Cortex, 31(2), 209–237. [DOI] [PubMed] [Google Scholar]
  35. Nozari N, Dell GS, & Schwartz MF (2011). Is comprehension necessary for error detection? A conflict-based account of monitoring in speech production. Cognitive Psychology, 63(1), 1–33. [DOI] [PMC free article] [PubMed] [Google Scholar]
  36. Postma A (2000). Detection of errors during speech production: A review of speech monitoring models. Cognition. [DOI] [PubMed] [Google Scholar]
  37. Price CJ (2010). The anatomy of language: A review of 100 fMRI studies published in 2009. Annals of the New York Academy of Sciences. [DOI] [PubMed] [Google Scholar]
  38. Rasmussen PM, Schmah T, Madsen KH, Lund TE, Yourganov G, Strother SC, & Hansen LK (2012). Visualization of nonlinear classification models in neuroimaging: Signed sensitivity maps. In BIOSIGNALS 2012 - Proceedings of the International Conference on Bio-Inspired Systems and Signal Processing. [Google Scholar]
  39. Roach A, Schwartz MF, Martin N, Grewal AS, & Brecher A (1996). The Philadelphia Naming Test: Scoring and rationale. Clinical Aphasiology, 24, 21–133. [Google Scholar]
  40. Romani C, Olson A, Semenza C, & Granà A (2002). Patterns of phonological errors as a function of a phonological versus an articulatory locus of impairment. Cortex, 38(4), 541–567. [DOI] [PubMed] [Google Scholar]
  41. Rorden C, Bonilha L, Fridriksson J, Bender B, & Karnath HO (2012). Age-specific CT and MRI templates for spatial normalization. NeuroImage, 61(4), 957–965. [DOI] [PMC free article] [PubMed] [Google Scholar]
  42. Schuchard J, Middleton EL, & Schwartz MF (2017). The timing of spontaneous detection and repair of naming errors in aphasia. Cortex, 93, 79–91. [DOI] [PMC free article] [PubMed] [Google Scholar]
  43. Schwartz MF, Faseyitan O, Kim J, & Coslett HB (2012). The dorsal stream contribution to phonological retrieval in object naming. Brain, 135(12), 3799–3814. [DOI] [PMC free article] [PubMed] [Google Scholar]
  44. Schwartz MF, Kimberg DY, Walker GM, Faseyitan O, Brecher A, Dell GS, & Coslett HB (2009). Anterior temporal involvement in semantic word retrieval: Voxel-based lesion-symptom mapping evidence from aphasia. Brain, 132(12), 3411–3427. [DOI] [PMC free article] [PubMed] [Google Scholar]
  45. Schwartz MF, Middleton EL, Brecher A, Gagliardi M, & Garvey K (2016). Does naming accuracy improve through self-monitoring of errors? Neuropsychologia, 84, 272–281. [DOI] [PMC free article] [PubMed] [Google Scholar]
  46. Schwartz MF, Wilshire CE, Gagnon DA, & Polansky M (2004). Origins of nonword phonological errors in aphasic picture naming. Cognitive Neuropsychology. [DOI] [PubMed] [Google Scholar]
  47. Stark JA (1988). Aspects of Automatic versus Controlled Processing, Monitoring, Metalinguistic Tasks, and Related Phenomena in Aphasia. [Google Scholar]
  48. Wilshire CE (2002). Where do aphasic phonological errors come from? Evidence from phoneme movement errors in picture naming. Aphasiology, 16(1–2), 169–197. [Google Scholar]
  49. Yeung Nicholas. (2015). Conflict Monitoring and Cognitive Control Oxford Handbooks Online Conflict Monitoring and Cognitive Control (Vol. 2). [Google Scholar]
  50. Yeung Nick, Botvinick MM, & Cohen JD (2004). The neural basis of error detection: conflict monitoring and the error-related negativity. Psychological Review, 111(4), 931–959. [DOI] [PubMed] [Google Scholar]
  51. Yourganov G, Fridriksson J, Rorden C, Gleichgerrcht E, & Bonilha L (2016). Multivariate Connectome-Based Symptom Mapping in Post-Stroke Patients: Networks Supporting Language and Speech. Journal of Neuroscience, 36(25), 6668–6679. [DOI] [PMC free article] [PubMed] [Google Scholar]
  52. Zhang Y, Kimberg DY, Coslett HB, Schwartz MF, & Wang Z (2014). Multivariate lesion-symptom mapping using support vector regression. Human Brain Mapping, 35(12), 5861–5876. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplemental Materials

Data Availability Statement

The data that support the findings of this study are available on request from the corresponding author. The data are not publicly available due to their containing information that could compromise the privacy of research participants. Source code that was used to conduct the SVR-LSM analyses in this study is available at https://github.com/atdemarco/svrlsmgui/.

RESOURCES