Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2018 Jan 1.
Published in final edited form as: Aphasiology. 2016 Sep 2;31(8):951–980. doi: 10.1080/02687038.2016.1225276

Rapid recovery from aphasia after infarction of Wernicke's area

Stephanie A Yagata a, Melodie Yen a,b, Angelica McCarron a, Alexa Bautista a, Genevieve Lamair-Orosco a, Stephen M Wilson a,b,c,
PMCID: PMC5642116  NIHMSID: NIHMS868165  PMID: 29051682

Abstract

Background

Aphasia following infarction of Wernicke's area typically resolves to some extent over time. The nature of this recovery process and its time course have not been characterized in detail, especially in the acute/subacute period.

Aims

The goal of this study was to document recovery after infarction of Wernicke's area in detail in the first 3 months after stroke. Specifically, we aimed to address two questions about language recovery. First, which impaired language domains improve over time, and which do not? Second, what is the time course of recovery?

Methods & Procedures

We used quantitative analysis of connected speech and a brief aphasia battery to document language recovery in two individuals with aphasia following infarction of the posterior superior temporal gyrus. Speech samples were acquired daily between 2 and 16 days post stroke, and also at 1 month and 3 months. Speech samples were transcribed and coded using the CHAT system, in order to quantify multiple language domains. A brief aphasia battery was also administered at a subset of five time points during the 3 months.

Outcomes & Results

Both patients showed substantial recovery of language function over this time period. Most, but not all, language domains showed improvements, including fluency, lexical access, phonological retrieval and encoding, and syntactic complexity. The time course of recovery was logarithmic, with the greatest gains taking place early in the course of recovery.

Conclusions

There is considerable potential for amelioration of language deficits when damage is relatively circumscribed to the posterior superior temporal gyrus. Quantitative analysis of connected speech samples proved to be an effective, albeit time-consuming, approach to tracking day-by-day recovery in the acute/subacute post-stroke period.

Keywords: stroke, aphasia, recovery, connected speech, Wernicke's area

Introduction

Wernicke's area, in the left posterior perisylvian region, is a key brain area for language (Wernicke, 1874). Indeed, many authors consider Wernicke's area to be the most fundamental and indispensable language region of the brain (Marie, 1906; Penfield & Roberts, 1959). Wernicke's area is generally defined as including the posterior superior temporal gyrus (STG), and depending on the definition, may include the adjacent supramarginal gyrus (SMG), middle temporal gyrus (MTG), and/or angular gyrus (Bogen & Bogen, 1976). In the acute stage, damage to Wernicke's area is associated with Wernicke's aphasia, characterized by impaired comprehension and repetition, and fluent paraphasic speech (Hillis et al., 2001).

Despite the apparently critical role of Wernicke's area in language processing, patients with infarction or resection of the posterior STG and adjacent structures often show considerable recovery of language function (Kertesz, Lau, & Polk, 1993; Laska, Hellblom, Murray, Kahan, & Von Arbin, 2001; Penfield & Roberts, 1959; Selnes, Knopman, Niccum, Rubens, & Larson, 1983; Selnes, Niccum, Knopman, & Rubens, 1984; Weiller et al., 1995). Recovery is highly variable (Kertesz & McCabe, 1977), and the extent of recovery depends on whether adjacent structures such as the SMG, angular gyrus and MTG are involved (Kertesz & Benson, 1970; Kertesz et al., 1993; Metter et al., 1990; Naeser, Helm-Estabrooks, Haas, Auerbach, & Srinivasan, 1987; Selnes et al., 1983).

The nature and time course of the recovery process after Wernicke's area infarction have not been characterized in detail. Previous studies have investigated just one or two language measures, such as single word comprehension (Selnes et al., 1984), sentence comprehension (Selnes et al., 1983), comprehension in general (Kertesz et al., 1993; Naeser et al., 1987), jargon in speech production (Kertesz & Benson, 1970) or overall aphasia severity (Kertesz et al., 1993; Kertesz & McCabe, 1977; Laska et al., 2001). Moreover, most studies have first assessed patients at an initial time point of 2 weeks to 1 month, meaning that recovery taking place prior to that point has not been documented. This early period constitutes an important missing piece of the recovery process, because in many cases, the greatest gains take place in the first few weeks after stroke (Basso, 1992; Pedersen, Stig Jørgensen, Nakayama, Raaschou, & Olsen, 1995).

However, there are practical challenges to studying recovery of language function in the immediate aftermath of a stroke. Patients' and their families' lives have just been turned upside down, and it is typically a traumatic and confusing time. Patients are transferred from emergency departments to intensive care units to acute care units. Some may be discharged home rather abruptly, within a day or two, if there are no physical deficits, while others are transferred to inpatient rehabilitation facilities. Patients with comprehension deficits are often not testable on any aphasia battery that requires following instructions and providing specific responses. Even if comprehension is spared, it is often not feasible to repeatedly administer a comprehensive aphasia battery such as the Boston Diagnostic Aphasia Examination (Goodglass, Kaplan, & Barresi, 2001). On the other hand, briefer aphasia screening tools that have been developed (e.g. Azuar et al., 2013; Flamand-Roze et al., 2011) do not quantify language in sufficient detail to fully capture the nature of recovery patterns.

An alternative approach is to assess recovery of language function through quantitative analysis of connected speech (Holland et al., 1985; MacWhinney, Fromm, Forbes, & Holland, 2011; Saffran, Berndt, & Schwartz, 1989; Wilson, Henry, et al., 2010). Connected speech samples can potentially be obtained through conversation at the bedside in the early days following a stroke, and quantitative analysis of connected speech can provide rich information about multiple language domains. In a single individual with global aphasia, Holland et al. (1985) used connected speech samples obtained daily to document recovery from global to anomic aphasia over a period of 2 weeks, suggesting that this approach can be an effective method for simultaneously collecting information about changes in multiple language domains at frequent time points. Several other studies have documented recovery from aphasia later in the subacute period based in part on longitudinal analyses of connected speech (Gandour, Marshall, Kim, & Neuburger, 1991; Kohn & Smith, 1994; Marshall, 1982).

The goal of the present study was to document recovery after infarction of Wernicke's area in detail in the first 3 months after stroke. Specifically, we aimed to address two questions about language recovery. First, which impaired language domains improve over time, and which do not? Second, what is the time course of recovery?

Methods

We studied two individuals with aphasia following acute left middle cerebral artery stroke. We collected 5-minute conversational samples of connected speech daily between 2 days and 16 days post stroke, and then at 1 month and 3 months post stroke. We transcribed and coded each sample using CHAT (MacWhinney, 2000; MacWhinney et al., 2011) in order to analyze the speech samples for multiple measures of expressive language. We supplemented our analyses of connected speech with a brief aphasia battery at a subset of five time points during the 3 months.

Participants

Two individuals with aphasia due to acute left hemisphere ischemic stroke to Wernicke's area participated in the study. Beginning in the summer of 2014, we began to screen all patients seen by the stroke service at the University of Arizona Medical Center, in order to identify individuals with new onset aphasia for longitudinal investigation. These two patients were the first two patients recruited, and both had focal damage to Wernicke's area. Both were native English speakers, and neither had a history of any pre-existing impairments of speech or language.

Connected speech samples were also obtained from 6 healthy native English speakers in their seventies (4 female; mean age 73.0 years; range 70–78 years), in order to establish normative ranges for connected speech measures. These participants were recruited after attending a talk on language and the brain at a community center. They were all right-handed, native speakers of English, and neurologically normal. Their scores on the Mini Mental State Examination (Folstein et al., 1975) ranged from 29 to 30.

The study was approved by the University of Arizona Institutional Review Board. All participants provided informed consent for the study, and were modestly compensated for their participation.

Participant JI

JI (not his real initials), a right-handed male, experienced a stroke in the summer of 2014, and was first approached 2 days after the stroke. He was 76 years old. Prior to his stroke, he had been working as a security guard. He had completed high school and 1 year of university-level education. He had no previous neurological history.

His family awoke to a commotion at 6:00 a.m., and observed JI fall. His family reported that he was unable to “find words”. He was brought to the emergency department by his family at 6:45 a.m. Nursing staff described “garbled speech upon arrival.” By 7:30 a.m., JI was reported to be following basic commands, but “only able to answer yes and no to questions.” He was taken for clinical MRI at approximately 8:00 a.m. At 11:45 a.m., his physician noted the following in regard to the onset of JI's symptoms: “I heard him say, ‘2 a.m.’, ‘5 a.m.’, and ‘2-1’; he is unable to verify time of onset by nodding or writing.” No other observations were made about JI's speech/language until the next afternoon, at which time his speech-language pathologist reported that he was reliably able to produce ‘yes’, ‘no’, and ‘let's try it’. In the morning, 2 days post stroke, the same speech-language pathologist noted improved expressive communication, but observed that he correctly named only one of the three items on his breakfast tray.

We first approached JI 2 days post stroke. We explained the study to him using simplified language (e.g. active sentence structures; high-frequency, early-acquired vocabulary), and we ascertained that he understood the required components of informed consent by talking with him about the study, answering his questions, and asking him questions to verify comprehension of key concepts, e.g. ‘Can you stop at any time?’. He consented to participate in the study, and a conversational speech sample and brief aphasia battery (see below) were acquired. At this time, JI presented with severe conduction aphasia, based on our clinical impression. He was fluent, with pervasive phonemic paraphasias, neologisms, and word-finding difficulties, and attempts to self-correct that were mostly unsuccessful. Confrontation naming and repetition were severely impaired. Single word comprehension was largely preserved, but sentence comprehension was poor, likely secondary to verbal working memory deficits.

JI remained in acute care for 4 days after which time he was transferred to an inpatient rehabilitation facility for 8 days. Following inpatient rehabilitation, he was discharged home. We obtained connected speech samples every day between 2 days and 16 days post stroke, and at 30 and 110 days post stroke. Brief aphasia batteries were obtained at five time points. A summary of the data acquired is shown in Table 1. JI continued to exhibit conduction aphasia over the course of the study, with severity decreasing markedly over time.

Table 1. Data acquired and connected speech sample characteristics.
Time (days post stroke) Location Duration of patient speech analyzed (s) Words Utterances Aphasia battery Imaging
JI 0 (stroke) Intensive care/MRI DWI
2 Acute care 99 134 37 yes
3 Acute care 253 312 68
4 Acute care 244 377 70
5 Acute care 275 692 106
6 Inpatient rehabilitation 248 433 92
7 Inpatient rehabilitation 320 562 106
8 Inpatient rehabilitation 246 394 78
9 Inpatient rehabilitation 331 622 91 yes
10 Inpatient rehabilitation 284 517 90
11 Inpatient rehabilitation 263 534 76
12 Inpatient rehabilitation 351 738 86
13 Inpatient rehabilitation 263 584 90
14 Home 271 511 58
15 Home 278 694 72
16 Home 261 585 79 yes
30 Outpatient MRI 219 459 62 yes T1-MPRAGE, DTI
110 Home 245 560 76 yes
VG -347 Intensive care/MRI DTI
-344 Acute care/MRI T1-SPACE
0 (stroke) Intensive care/MRI DWI, FLAIR
2 Acute care 33 32 9 yes
3 Acute care 220 227 38
4 Acute care 256 266 34
5 Acute care 260 168 31
6 Acute care 217 210 35
8 Inpatient rehabilitation 164 173 33
9 Inpatient rehabilitation 196 228 37 yes
10 Inpatient rehabilitation 256 251 35
11 Inpatient rehabilitation 253 126 22
12 Inpatient rehabilitation 261 283 41
13 Inpatient rehabilitation 264 210 24
14 Inpatient rehabilitation 247 349 45
15 Inpatient rehabilitation 240 276 51
16 Inpatient rehabilitation 152 265 32 yes
41 Home 158 333 55 yes
92 Home 218 487 81 yes

JI did not receive speech-language therapy for aphasia during acute care. However, he received treatment for aphasia from a speech-language pathologist during inpatient rehabilitation one to two times daily, targeting word retrieval strategies. Upon discharge home, he did not receive continued speech-language therapy.

On the morning of JI's stroke, MR imaging was acquired according to a standard stroke protocol, including DWI, FLAIR, T1-weighted and GRE. Arterial spin labeling perfusion imaging was attempted, but was not completed due to lack of patient cooperation. At JI's 1-month follow-up visit, we acquired three-dimensional T1-weighted structural images, FLAIR and DTI on a Siemens 3T Skyra scanner.

All available images were coregistered to one another using SPM5 (Friston, Ashburner, Kiebel, Nichols, & Penny, 2007). The T1-MPRAGE image was transformed to MNI space using Unified Segmentation in SPM5, and the same transformation was applied to all other images. The lesion was drawn manually using ITK-SNAP 3.2.0 (Yushkevich et al., 2006), based on acute DWI (mean diffusivity, Figure 1A and apparent diffusion coefficient, Figure 1B) and 1-month FLAIR images (Figure 1C). The arcuate fasciculus was tracked using FSL 5.0 (Behrens, Berg, Jbabdi, Rushworth, & Woolrich, 2007), based on two masks: one in the anterior-posteriorly oriented fibers of the superior longitudinal fasciculus, and the other in the dorsal-ventrally oriented fibers behind the Sylvian fissure (Figure 1D), following previously described procedures (Galantucci et al., 2011).

Figure 1.

Figure 1

JI neuroimaging and lesion mapping. (A) lesion outlined on acute DWI (mean diffusivity); (B) lesion outlined on acute DWI (apparent diffusion coefficient); (C) lesion outlined on FLAIR at 1 month post-stroke time point; (D) arcuate fasciculus identified on DTI at 1 month post-stroke time point; (E) lesion shown on a series of sagittal images; (F) lesion shown on sagittal, coronal, and axial images with respect to the arcuate fasciculus, which is entirely disconnected.

JI's lesion, shown in a series of sagittal images in Figure 1E, was primarily concentrated in the posterior STG extending into the SMG and the angular gyrus. The lesion extended to the posterior insula. The ascending part of the posterior arcuate fasciculus was entirely destroyed by the lesion (Figure 1F).

Participant VG

VG (not her real initials), a right-handed female, experienced a stroke in the summer of 2014, and was first approached 1 day after the stroke. She was 85 years old. VG was a retired operating room nurse, and prior to her stroke, she had been an active member in her local senior center. Her neurological history was significant for a previous right hemisphere stroke in the basal ganglia and centrum semiovale in the summer of 2013. This previous stroke had not resulted in any speech or language deficits.

VG presented to the emergency department at approximately 6:30 a.m. on the day of her stroke after a neighbor found her at home “unable to speak clearly.” The same neighbor had seen her the previous afternoon at which time she had not shown any signs or symptoms of stroke. Upon admission, nursing staff described her speech as “garbled.” Around 7:00 a.m., a physician noted that she was able to answer some yes/no questions, but that she was not a reliable historian. VG was taken for clinical MRI at approximately 9:00 a.m. Nursing staff noted that at 10:30 a.m. she was “verbalizing understanding,” but around 4:00 p.m., her case manager reported that she was “unable to verbalize.”

The next morning, one of the hospital's speech-language pathologists described VG as “aphasic, with comprehension appearing more intact than expression,” and noted that her language was characterized by “frequent paraphasias and perseverations.” The speech-language pathologist also noted some dysarthria. We approached VG later that day, at which time she was largely non-verbal, answering only ‘yes’ or ‘no’ to questions. At that point, we judged her comprehension too impaired for her to provide informed consent.

We approached her again the next day (2 days post stroke), at which point her comprehension had improved somewhat. As described above for JI, we explained the study to VG and talked with her about it until we were satisfied that she understood the required components of informed consent. She provided consent, and we acquired a conversational sample, and a brief aphasia battery (see below). At this time, VG presented with moderate Wernicke's aphasia, based on our clinical impression. She was fluent, but her connected speech was characterized by pervasive word-finding difficulties and paragrammatism, some jargon, and occasional semantic and phonemic paraphasias. Confrontation naming and repetition were severely impaired. Her comprehension was impaired at the single word and sentence level. She also exhibited moderate dysarthria of mixed spastic/hypokinetic type.

VG remained in acute care for 6 days following her stroke, and was transferred to an inpatient rehabilitation facility on day 7. She remained in inpatient rehabilitation for 16 days, at which time she was discharged home. We obtained connected speech samples every day between 2 days and 16 days post, except for day 7 on which she was exhausted due to the transfer to the rehabilitation facility. We also obtained samples at 41 and 92 days post stroke. Brief aphasia batteries were obtained at five time points. A summary of the data acquired is shown in Table 1. Over the course of the study, VG's language impairment evolved from Wernicke's aphasia to conduction aphasia, as her comprehension and word finding improved. Her dysarthria gradually reduced in severity from moderate to mild-moderate by 3 months post stroke.

VG was evaluated but not treated for aphasia during acute care, and she received speech-language therapy for aphasia one to two times daily during her period of inpatient rehabilitation, focusing on word retrieval strategies. After she returned home, VG received occupational therapy, but no speech-language therapy.

On the morning of her stroke, MR images were acquired according to a standard stroke protocol, including DWI, FLAIR, T1-weighted, T2-weighted and GRE. Dynamic susceptibility contrast perfusion imaging was performed but was non-diagnostic due to a technical issue. Although there were no three-dimensional structural images or DTI acquired at this time, these images were available from approximately 1 year prior, having been acquired after the previous right subcortical stroke. The T1-weighted structural image was of low quality due to motion artifact, however it was adequate for visualization of key structures.

The images were processed as described for JI, except that the acute FLAIR was used along with DWI for lesion delineation, since the acute imaging was acquired sufficiently late that the lesion appeared hyperintense on FLAIR. Lesion delineation is shown in Figure 2A-C and fiber tracking in Figure 2D.

Figure 2.

Figure 2

VG neuroimaging and lesion mapping. (A) lesion outlined on acute DWI (mean diffusivity); (B) lesion outlined on acute DWI (apparent diffusion coefficient); (C) lesion outlined on acute FLAIR; (D) arcuate fasciculus identified on DTI; (E) lesion shown on a series of sagittal images; (F) lesion shown on sagittal, coronal, and axial images with respect to the arcuate fasciculus, which is entirely disconnected.

VG's lesion, shown in a series of sagittal slices in Figure 2E, encompassed much of the left posterior STG and posterior superior temporal sulcus (STS) as well as portions of the MTG and SMG. The lesion extended medially, effectively dissecting the ascending part of the posterior arcuate fasciculus (Figure 1F). There were also lesions to the left basal ganglia and posterior insula, which extended into external capsule fibers. In combination with the previous lesion of the right basal ganglia, these lesions were presumably responsible for VG's motor speech deficits.

Connected speech samples

Elicitation and recording of connected speech samples

Connected speech samples were elicited by the first author (SAY, a second year masters student in Speech-Language Pathology) based on predetermined conversational prompts (Table 2). A different prompt was used each session, in order to elicit more natural conversational speech (rather than repeating the same questions, day after day). When necessary, more than one prompt was used in a session.

Table 2. Prompts for elicitation of connected speech samples.
‘Tell me about…’ what happened when you had your stroke
what you did yesterday
the best trip you ever took
the worst trip you ever took
your favorite holiday as a child
a happy childhood memory
your worst childhood memory
how you met your husband/wife/partner
when you got married
when you had your first child
your first job
when you retired
a time you were really scared/embarrassed/angry
Example prompts, if no response in approximately 10 seconds: ‘Do you remember a trip you took?’
‘Tell me about your favorite Thanksgiving/Christmas.’
‘Tell me about your wedding.’

We aimed to collect at least 5 minutes of conversation, although a few sessions, including both patients' 2 days post time points, and some of the later time points that included aphasia batteries, were somewhat shorter than this because we wanted to minimize our imposition on the patients. When more than 5 minutes were acquired, only the first 5 minutes were analyzed. From this 5 minute sample (or less, if less than 5 minutes was acquired), the sample duration (Table 1) was calculated as the total time that the patient was speaking (including pauses). Excluded from the sample duration were examiner speech, pauses between examiner speech and patient speech, patients' direct responses to questions, and any interruptions.

All conversations were recorded using a Sanken lavalier microphone (COS-11D) and a Canon camcorder (Vixia HF S20).

Transcription and coding of connected speech samples

Transcription and coding of the connected speech samples followed the CHAT system (MacWhinney, 2000, 2012), including newer codes that were introduced for quantifying aphasic speech (MacWhinney et al., 2011). CHAT provides a standardized way to represent naturalistic speech samples in aphasia with the means to capture phonological, lexical and morphosyntactic disturbances. Speech samples were transcribed using EUDICO Linguistic Annotator (ELAN; Wittenburg, Brugman, Russel, Klassmann, & Sloetjes, 2006) in order to encode precise timing information. All speech samples were transcribed in standard English orthography except for the use of International Phonetic Alphabet (IPA) to transcribe phonemic paraphasias and neologisms. Each sample was transcribed and coded by either SAY, MY (a first year Ph.D. student in Linguistics) or GL-O (a fourth year undergraduate major in Speech, Language and Hearing Sciences), all of whom were trained in the lab of SMW, a researcher with experience in the analysis of connected speech in aphasia (Wilson, Henry, et al., 2010). All samples were then comprehensively checked and edited by SAY to maximize consistency of transcription and coding.

Utterance boundaries were determined based on the principles described by Saffran et al. (1989). Errors were coded at the word level and/or at the utterance level. Word-level errors were classified as phonological errors, semantic errors, neologisms, formal lexical device errors, morphological errors, or omissions (MacWhinney, 2012). Word-level errors were counted regardless of whether or not they were retraced and/or corrected, and if errors were repeated, they were counted each time. Five utterance-level error codes defined in the CHAT system were used to code grammatical errors (agrammatic or paragrammatic), empty speech, circumlocution, jargon and perseverative utterances (the last two of these categories were infrequent so were not used in our analyses). We defined three additional codes at the utterance level: semantically anomalous utterances, non-sentence utterances (e.g. isolated noun phrases), and embeddings (see definitions below). Utterances could meet the criteria for multiple utterance-level codes. To avoid double-counting errors, the utterance-level codes for grammatical errors and semantically anomalous utterances were not used when the errors in question could be captured by word-level codes.

After transcription and coding, each morpheme was tagged for part of speech using the program mor (MacWhinney, 2000), then manually disambiguated.

We analyzed the coded and tagged transcriptions by calculating measures (described below) to capture change over time in four broad language domains: (1) speech rate and speech sound errors; (2) lexical content; and (3) morphosyntax and complexity; (4) disruptions to fluency. An in-house MATLAB program was used to derive all measures from the coded transcripts, and to carry out statistical analyses.

Speech rate and speech sound errors

Speech rate in words per minute was calculated as the total number of words produced divided by the time spent speaking. All real words and neologisms were counted, whereas fillers, false starts, unintelligible speech, words that were subsequently retraced, and direct responses to examiner questions were excluded. The time spent speaking was the duration of the speech sample, minus examiner speech, pauses between examiner speech and patient speech, direct responses to examiner questions, and interruptions.

Phonological errors were classified as phonemic paraphasias or neologisms. Phonemic paraphasias were defined as a substitution, omission, or deletion of the onset, nucleus, or coda in one syllable of the target word (MacWhinney, 2012). Neologisms were defined as words with unknown targets, or when the target was known, words that contained more than one substitution, omission, or deletion. In calculating the measure of phonological errors, neologisms were each weighted as two as they contained at least two errors. This measure, along with many other measures described below, was calculated per hundred words in order to normalize for sample length.

One patient, VG, had significant dysarthria that changed across the 3-month period of the study. Dysarthric errors were not individually coded in the samples. Instead, to capture change in motor speech over time, five graduate students in speech-language pathology listened to each of VG's speech samples presented in a pseudorandomized order and rated them for dysarthric features. The students were unfamiliar with VG prior to rating the samples. They were presented only with audio recordings, and not with video, in order to avoid visual cues as to the patient's stage of recovery. The five students rated each sample for seven features: prosody, rate, resonance, articulation, voice quality, respiration, and overall intelligibility. Each feature was rated on a seven-point scale: 0 = no impairment; 1 = mild; 2 = mild-moderate; 3 = moderate; 4 = moderate-severe; 5 = severe; 6 = mute. The features and rating scale were loosely based on Darley et al. (1969) and Duffy (2013). All students had completed coursework on motor speech disorders in which numerous dysarthric speech samples had been rated on similar scales, and all had some experience evaluating clients with motor speech disorders. A mean dysarthria score for each sample was computed by averaging scores across all five students and all seven features.

Lexical content

A composite measure of impaired lexical access was calculated by adding the counts of word-level semantic errors, neologisms with unknown targets, incomplete utterances (those that trailed off, almost always due to word-finding problems), utterances containing empty speech, utterances containing circumlocution, and semantically anomalous utterances. The utterance-level code for semantically anomalous utterances was used when utterances were semantically inappropriate for the context, but where the semantic anomaly could not be pinned down to any word-level problem.

The proportion of closed class words was calculated by dividing the number of closed class words by the total number of open and closed class words. Similarly, the proportion of verbs was calculated by dividing the number of verbs by the number of verbs and nouns. These proportions are indirect measures of agrammatism (which results in reduced proportions of closed class words and verbs) and empty speech (which results in elevated proportions of closed class words and verbs) (Wilson, Henry, et al., 2010).

Morphosyntax and complexity

Mean length of utterance in words, a commonly used metric of syntactic complexity, was calculated by dividing the total number of words in each sample (excluding fillers, false starts, unintelligible speech, and retraced words) by the number of utterances.

Bound morphemes were counted as a measure of morphological complexity, including inflectional and derivational morphemes. Incorrectly used bound morphemes were included in the count as they demonstrated attempts to use morphology.

Utterances containing embeddings were coded as a measure of syntactic complexity. Embeddings were required to contain a subject or a finite verb form. If an utterance contained multiple embeddings, each was counted. Embeddings were counted even if utterances were not grammatical, but the defining criteria for embedding needed to be clearly present.

A composite measure of morphosyntactic errors was calculated by adding the counts of word-level morphological errors, errors in formal lexical devices, omitted words (where these could be determined with reasonable confidence), utterances that were not complete sentences (e.g. noun phrases, verb phrases, etc.), and agrammatic or paragrammatic utterances. Because the utterance-level code for agrammatic/paragrammatic speech was used only when utterances contained more than one word-level morphosyntactic error, it was given double weight in the composite measure.

Disruptions to fluency

In this category, we included three further phenomena that contribute to an overall impression of non-fluent speech: filled pauses (e.g. ‘uh’, ‘um’, and ‘er’), false starts and retracings. False starts were defined as partial words, i.e. words that were abandoned, usually after just one or a few phonemes had been produced. Retracings were defined as sequences of one or more complete words (i.e. not solely a filler or a false start), which were made redundant by subsequent repetitions, amendments, elaborations or alternative expressions.

Brief aphasia battery

At a subset of five time points (Table 1), each patient's speech and language was further assessed using a brief in-house aphasia battery, which was designed to be quicker to administer than commonly used batteries and so more appropriate for the acute and subacute time points.

The battery quantified expressive and receptive function through five language tasks: (1) confrontation naming; (2) single word comprehension; (3) sentence comprehension; (4) repetition of words and sentences; and (5) reading aloud of words and sentences. The confrontation naming portion consisted of 12 items, varying in word length and frequency, with high name agreement and monomorphemic targets. Pictures were colorized versions of Snodgrass and Vanderwert's line drawings (Rossion & Pourtios, 2004). The single word comprehension task comprised 16 items, 4-alternative forced choice auditory word-picture matching. There were 12 items with semantic, phonemic, and unrelated foils, and 4 difficult low-frequency items (e.g., ‘asparagus’, ‘artichoke’) with 3 closely related semantic foils each. The sentence comprehension task consisted of 12 items, with 4-alternative forced-choice sentence-picture matching. The stimuli were similar to those created for a previous study (Wilson, Dronkers, et al., 2010), and varied in terms of length and syntactic complexity. The repetition and reading tasks involved single words of varying length, frequency, regularity, and morphological complexity (40%), sentences of varying length and word frequency including items loaded on function words (40%); and pseudowords of varying length (20%).

The battery was administered by SAY, AM (a third year undergraduate major in Speech, Language and Hearing Sciences) or AB (a fourth year undergraduate major in Speech, Language and Hearing Sciences), all of whom were trained by SMW, who designed the battery. All sessions were videotaped, then transcribed and scored by AM or AB. Correct responses scored one point each, except where they were delayed, self-corrected, or where repetitions were requested, in which case they scored half a point (as in the Comprehensive Aphasia Test; Swinburn, Porter, & Howard, 2004).

Statistical analysis

To identify overall patterns of change and the relationships of the language variables to one another, principal components analysis (PCA) was performed based on correlations (i.e. the variables were normalized, because they had different scales). The dysarthria variable was not included for VG in order to make her PCA results comparable to those for JI (but including the dysarthria variable did not significantly change the results).

For each language variable described above, and for each patient, we assessed whether the variable showed a significant change over time. Linear models of language measures by time generally resulted in skewed residuals supporting the necessity of log-transforming the independent variable. Moreover, previous research has shown that recovery from aphasia in the acute/subacute stages after a stroke is characterized by a “decelerating” curve, with the greatest gains made early, and the pace of recovery slowing over time (Basso, 1992; Pedersen et al., 1995). Therefore we assumed a logarithmic time course, and for each variable in each patient we fit the equation y = a.log(t) + b, where y is the language variable in question, t is time (days past stroke), and a and b are coefficients.

Since the direction of change was assumed to be positive for measures associated with successful language use (e.g., mean length of utterance), and negative for measures measuring erroneous language use (e.g., phonological errors), we computed one-tailed P-values accordingly for most measures. Two-tailed P-values were computed for the proportion of closed class words, and the proportion of verbs, since these quantities can be perturbed in either direction in aphasia (Wilson, Henry, et al., 2010).

Because we investigated the time course of numerous language variables in two different patients, the issue of multiple comparisons must be considered. For JI, 11 of 17 statistical tests were significant at p ≤ 0.05. With a false discovery rate of q = 0.1 (Benjamini & Hochberg, 1995), all 11 would remain significant. For VG, 9 of 18 statistical tests were significant at p ≤ 0.05. With a false discovery rate of q = 0.1, all 9 would remain significant. Because all p-values ≤ 0.05 remain significant, for the sake of simplicity, we present uncorrected p-values for the remainder of the paper.

Correlation coefficients are described as little to none (r < 0.1), weak (0.1 ≤ r < 0.3), moderate (0.3 ≤ r < 0.5) or strong (r ≥ 0.5), independently of statistical significance (Cohen, 1988).

Results

Principal components analysis

All changes over time in all language measures are summarized in Table 3.

Table 3. Changes in language measures.

Measure Patient

JI VG
r p r p

First principal component 0.81 < 0.0001 *** 0.66 0.0029 **
Speech rate and speech sound errors
 Words per minute 0.64 0.0030 ** 0.73 0.0006 ***
 Phonological errors (phw) –0.72 0.0005 *** –0.39 0.067
 Dysarthria N/A –0.55 0.0129 *
Lexical content
 Closed class words (proportion) –0.77 0.0003 *** 0.52 0.037 *
 Verbs (proportion) –0.49 0.048 * 0.36 0.17
 Impaired lexical access (phw) –0.83 < 0.0001 *** –0.43 0.048
Morphosyntax and complexity
 Mean length of utterance (words) 0.65 0.0024 ** 0.28 0.15
 Bound morphemes (phw) 0.37 0.070 0.19 0.24
 Embeddings (phw) 0.60 0.0058 ** 0.54 0.0153 *
 Morphosyntactic errors (phw) –0.08 0.38 –0.42 0.0537
Other disruptions to fluency
 Filled pauses (phw) –0.34 0.091 –0.47 0.0321 *
 False starts (phw) –0.22 0.20 –0.32 0.11
 Retracings (phw) –0.58 0.007 ** –0.29 0.14
Aphasia battery
 Confrontation naming 0.96 0.0045 ** 0.93 0.012 *
 Single word comprehension 0.71 0.090 0.34 0.29
 Sentence comprehension 0.51 0.19 0.85 0.034 *
 Repetition 0.98 0.0014 ** 0.93 0.010 *
 Reading aloud 0.94 0.0079 ** 0.73 0.082
*

p ≤ 0.05,

**

p ≤ 0.01,

***

p ≤ 0.001.

phw = per hundred words.

PCA was performed for each patient to identify overall patterns of change and the relationships between the language variables. For each patient, scree plots revealed that the first principal component accounted for a large portion of the variance: 48% in JI and 40% in VG.

For each patient, the first principal component tracked recovery over time with a logarithmic time course (JI: strong correlation between first principal component score and log-transformed time, r = 0.81, p < 0.0001, Figure 3A; VG: strong correlation, r = 0.66, p = 0.0029, Figure 3B). The factor loadings of these first principal components were similar for the 2 patients, with a few notable exceptions (Figure 3C). In general, the first principal components were positively weighted on variables associated with successful language use and negatively weighted on variables measuring erroneous language use. The exceptions to this general pattern were the proportions of closed class words and verbs, which changed in opposite directions in the 2 patients, and morphosyntactic errors, which changed over time only in VG and not in JI.

Figure 3.

Figure 3

Principal components analysis (PCA) of language measures over time. (A) Change over time in the first principal component in JI. (B) Change over time in the first principal component in VG. (C) Loading of 12 language variables on the first principal component for each of the 2 patients.

The second principal components did not change over time in either patient (JI: p = 0.57; VG: p = 0.96), suggesting that only the first principal components were meaningful.

Speech rate and speech sound errors

Speech rate in words per minute increased over time for both patients (JI: strong correlation, r = 0.64, p = 0.0030, Figure 4A; VG: strong correlation, r = 0.73, p = 0.0006, Figure 4B). These increases in words per minute reflected more rapid production of individual words, decreased pauses within and between utterances, and especially for JI, decreases in the number of retracings (see below) across the recovery period. By 3 months post stroke, both patients' speech rates were within in the control range, which is indicated by the gray shaded area in Figure 4 and subsequent figures.

Figure 4.

Figure 4

Change over time in speech rate and speech sound errors. (A, B) Words per minute. Data for the 6 control participants are shown (dotted gray line = mean; light gray shaded area = one standard deviation). Note that the x axis (time post stroke) is shown with a logarithmic scale. (C, D) Phonological errors. phw = per hundred words. Note that controls made negligible numbers of phonological errors. (E) JI had no motor speech deficit, so dysarthria was not rated. (F) Dysarthria in VG.

The prevalence of phonological errors decreased over time in both patients (JI: strong correlation, r = –0.72, p = 0.0005, Figure 4C; VG: moderate marginally significant correlation, r = –0.39, p = 0.067, Figure 4D). Examples are provided below. Earlier samples were characterized by the presence of pervasive phonemic paraphasias and neologisms (1a-c, 2a,b), which reduced in frequency and in terms of deviance from intended targets over the course of recovery (1d,e, 2c-e). The control participants made negligible numbers of phonological errors (2 of the 6 controls made a single error each).

In this paper, the following conventions apply to all examples: (i) examples are brief excerpts from each patient's speech samples, selected as representative of the patterns being described; (ii) phonological errors are transcribed in IPA and contained in brackets; (iii) intended targets are represented in parentheses with an equal sign; (iv) omitted words are indicated by a Ø followed by the target word; (v) retraced sequences are surrounded by angled brackets; (vi) false starts are transcribed in IPA with a hyphen to indicate that the word was incomplete; (vii) pauses greater than 1 second are indicated with (.); (viii) incomplete utterances are indicated with ellipses; (ix) participant initials, days post stroke, and contextual information are included in parentheses following each example.

    1. I fell out went this way and I [slɪmɪɾəd] (=slipped) a θ- [θiɔɹəbət] (=unknown). (JI, 2 days post, about his stroke)
    2. <I [dɪdəl] (=did)> I uh uh did a lot of uh oh work with the uh [neɪmɚl̩] (=unknown). and then we'd [teɪ] (=take) over to here and over here. and we'd take this. and there we got uh all uh [pʊɾeɪ] (=unknown) and [pəɾeɪɹə] (=unknown). (JI, 3 days post, about his first job)
    3. yeah well <[suplub] (=seafood) I just like> I like uh sipl- uh <[sipɹplud] (=seafood)> pɹu- pɹub- [sipɹrub] (=seafood). (JI, 8 days post, about his favorite food)
    4. and everything and my wipe (=wife) s- she locked (=liked) all the people and I locked (=liked) the guy <it just seemed> I think that was probably the best one. (JI, 11 days post, about his favorite job)
    5. <and so> and I wanna go <and uh for more> for eight more [smʌnθs] (=months) for <the> [baɪ] (=my) echo. (JI, 110 days post, about a doctor's appointment)
    1. what do you wanna know? well (.) not sure. well I do get <a> (.) a (.) [kæ̂səm̩] (=unknown) (.) um (.) canes (=unknown). mate (=unknown) uh (.) heart s- (.) um (.) b- b- <[blæs] (=unknown)> [blæsəz] (=unknown) a bunch of stuff. (VG, 2 days post, about her stroke)
    2. <you> you can (.) make a (.) k- <[kæ̂ʔʌ] (=unknown)> (.) [kænləs] (=unknown) (.) əwə- æ̃wɪ- and this (.) outfit (=unknown). (VG, 5 days post, about her activities)
    3. I s- sang. (.) I was a (.) æl- <[ældoʊdɚ] (=alto)> [ælddoʊɚ] (=alto). (VG, 11 days post, about her high school life)
    4. and so I <had a> uh (.) um (.) um (.) had some [taɪn] (=time) away to g- get away from work. (.) so I caught <a a a <[bɜ˞s] (=bus)> a b-> a bus. (VG, 15 days post, about a trip to Canada)
    5. but I'm not a therapist. so how do I know what that's all about? (.) and I just [əksɛkt] (=accept) it and go on. (VG, 41 days post, about her therapy)

Dysarthria was a notable feature of VG's speech. Based on the evaluations of five raters blinded to time point, the severity of her dysarthria decreased over time (strong correlation, r = −0.55, p = 0.013, Figure 4F). This overall improvement was driven by significant improvements in prosody (r = −0.74, p = 0.0005), rate (r = −0.59, p = 0.0076), respiration (r = −0.58, p = 0.0092), and a marginal improvement in intelligibility (r = −0.34, p = 0.087), whereas articulation (r = −0.04, p = 0.44) and voice quality (r = −0.26, p = 0.16) showed no evidence of change over time. JI had no motor speech deficits, so dysarthria was not rated for him.

Lexical content

Lexical access improved over time for both patients, as evidenced by reductions in the composite measure of impaired lexical access (JI: strong correlation, r = –0.83, p < 0.0001, Figure 5A; VG: moderate correlation, r = –0.43, p = 0.048, Figure 5B). This measure summed across five phenomena that are indicative of impaired lexical access, all five of which were evident in both patients: word-level semantic paraphasias (e.g. ‘caught’ for ‘picked’ in 3c, ‘Christmas’ for ‘fourth of July’ in 4a), incomplete utterances (e.g. ‘you can hit the red’ in (3d) where the final noun ‘button’ is missing), empty speech (3a-c, 4b), circumlocution (4b), and semantically anomalous utterances (4c). Imperfect lexical access is not necessarily abnormal: 5 of the 6 healthy control participants had at least one instance of impaired lexical access in their samples. However, both patients were still outside the range of healthy controls by 3 months post stroke.

Figure 5.

Figure 5

Change over time in measures of lexical content. (A, B) Composite measure of impaired lexical access, reflecting word-level semantic paraphasias, neologisms with unknown targets, incomplete utterances, empty speech, circumlocution and semantically anomalous utterances. phw = per hundred words. (C, D) Proportion of closed class words (out of all words). (E, F) Proportion of verbs (out of verbs and nouns).

  • (3)
    1. and then he helped him um (.) <[pi] (=put) the uh the uh uh> put the um uh (.) thing together. and they uh… oh heavens so then… okay so then we got <all k-> all [gæd] (=good). (JI, 4 days post, about getting ready for his wedding)
    2. so that was that. but that said but that was when he was not the first time [ðæs] (=this) one but the second time. (JI, 7 days post, about a scary childhood memory)
    3. okay we got those there. and then <we we d-> we <uh kind uh kit (=picked) uh caught (=picked) uh> picked up some supplies and some food. and just [beʔ] (=unknown). we did uh a lot there. (JI, 14 days post, about running errands)
    4. just give it. you can hit the red… here I'll get it. if you don't know how to work it I'll get it. I just turned it down. I just wasn't even paying attention to it but if you want it there we go. (JI, 110 days post, about turning off the television)
  • (4)
    1. we had to wait for the (.) Christmas (=fourth of July) (.) and then we buried Ø after. (VG, 3 days post, about her mother's funeral)
    2. and the red ones is every now and then. but we have a lot of them. yeah. and I also got a uh <oh the red one (.) bou- [boʊgəngeɪjə] (=bougainvillea)> [geɪjə] (=bougainvillea) whatever Ø <call> call the things. (VG, 14 days post, about flowers in her garden)
    3. uh the Dad uh knew we were having trouble with that house down there. (.) what happened is when they uh came he put the name in <his> d- his outfit (=unknown). (VG, 92 days post, about building a house)

There were only two measures that changed in opposite directions in the 2 patients: the proportion of closed class words, and the proportion of verbs (out of nouns and verbs). In JI, both of these measures decreased over time and approached the healthy control range by 3 months post stroke (closed class words: strong correlation, r = −0.77, p = 0.0003, Figure 5C; verbs: moderate correlation, r = −0.49, p = 0.048, Figure 5E), as his speech became progressively less empty (see examples in (3) above).

In contrast, in VG, both of these measures increased over time (closed class words: strong correlation, r = 0.52, p = 0.037, Figure 5D; proportion of verbs: moderate but non-significant correlation, r = 0.36, p = 0.17, Figure 5F). The significance of the first of these measures was driven by the first time point, during which VG's speech output was minimal and agrammatic, resulting in a below-normal proportion of closed class words. When the first time point was excluded, there was no significant subsequent change in VG's proportion of closed class words (r = 0.31, p = 0.27). The trends to increased proportions of closed class words and verbs should not be interpreted as indicative of recovery, since these measures were higher than for control participants.

Morphosyntax and complexity

The morphological and syntactic complexity of both patients' utterances increased over time, as shown in the examples in (5) and (6). This was reflected in positive trajectories of three variables. Words per utterance increased in both patients (JI: strong correlation, r = 0.65, p = 0.0024, Figure 6A, VG: weak non-significant correlation, r = 0.28, p = 0.15, Figure 6B). The number of bound morphemes per hundred words, a measure of morphological complexity, increased in both patients (JI: moderate marginally significant correlation, r = 0.37, p = 0.070, Figure 6C; VG: weak non-significant correlation, r = 0.19, p = 0.24, Figure 6D). The number of embeddings per hundred words, a measure of syntactic complexity, increased in both patients (JI: strong correlation, r = 0.60, p = 0.0059, Figure 6E; VG: strong correlation, r = 0.54, p = 0.015, Figure 6F). For both patients, all three of these measures approached or were in the control range by 3 months post stroke.

Figure 6.

Figure 6

Change over time in measures of morphosyntax and complexity. (A, B) Mean length of utterance, in words. (C, D) Bound morphemes. phw = per hundred words. (E, F) Embeddings. (G, H) Morphosyntactic errors, including word-level morphological errors, errors in formal lexical devices, omitted words, isolated phrases, and agrammatic or paragrammatic utterances.

  • (5)
    1. what do they call these thing (=things)? down here. and she got that. and uh Ø (=I) think that's it. (JI, 3 days post, about getting extra blankets)
    2. uh <then> then this little doggy. he's run (=running) around all the park and every (=everything) for four days. and then we f- let him in. (JI, 6 days post, about getting his third dog)
    3. <it's uh> it seem it's been that every (=ever) since you know. except when I first had the stroke you know then I couldn't even say one word. (JI, 16 days post, about word finding)
    4. but I Ø (=am) doing quite well at work because… b- and the people at work think I'm doing okay. (JI, 110 days post, about returning to work)
    5. oh what have I been doing other than uh <doggie> walking the dogs? m- uh we've done that. um uh I think basically we've just been getting all the um um medical stuff taken care of. (JI, 110 days post, about what he has been doing recently)
  • (6)
    1. he was gonna surgery (.) or (.) uh a doctor. but <he> (.) he change it. (VG, 4 days post, about her husband)
    2. I've had a lot of uh (.) s- surgeon (=surgeries). (VG, 6 days post, about being an operating nurse)
    3. The w- birds was (=were) <a color> a color. and we had those whatever it w-… anyway we had a birds coming out of our ear (=ears). But <they are> they're interested (=interesting). <I had> he had build (=built) a h- h- home for the birds. (VG, 10 days post, about past pets)
    4. she said eventually it would come back (.) in time. (.) so I don't know. I'm in hopes it will. (VG, 41 days post, about her language improving; note: “in hopes” appears to be a dialectal feature; VG grew up in Louisiana)
    5. so they lived together for twenty-seven years. everything was calm. then he got married. (VG, 92 days post, about her parents and brother)

While the morphological and syntactic complexity of JI's utterances increased, his prevalence of morphosyntactic errors remained essentially unchanged over time, and was higher than the control range (little to no correlation, r = –0.08, p = 0.38, Figure 6G). Throughout the time period of our study, he omitted some function words (subject in 5a, auxiliary verb in 5d) and bound morphemes (plural in 5a, progressive in 5b, third person singular suffix in 5c), and some of his utterances were paragrammatic (e.g. 5b,c).

In contrast, VG made progressively fewer morphosyntactic errors over the course of her recovery, approaching the control range (moderate marginally significant correlation, r = −0.42, p = 0.054, Figure 6H). In the first few days after her stroke, her speech was sometimes agrammatic (6a), before taking on a paragrammatic character (6b, 6c), which then improved over time (6d, 6e).

Disruptions to fluency

Filled pauses, false starts, and retracings, phenomena that contribute to an overall impression of reduced speech fluency, all decreased in prevalence over time in both patients (filled pauses in JI: moderate marginally significant correlation, r = −0.34, p = 0.091, Figure 7A; filled pauses in VG: moderate correlation, r = −0.47, p = 0.032, Figure 7B; false starts in JI: weak non-significant correlation, r = −0.22, p = 0.20, Figure 7C; false starts in VG: moderate but non-significant correlation, r = −0.32, p = 0.11, Figure 7D; retracings in JI: strong correlation, r = −0.58, p = 0.007, Figure 7E; retracings in VG: weak non-significant correlation, r = −0.29, p = 0.14, Figure 7F). Examples are shown in (7) and (8). By 3 months post stroke, both patients continued to produce more false starts than controls, but approached controls in numbers of retracings, and were in the normal range for filled pauses.

Figure 7.

Figure 7

Change over time in disruptions to fluency. (A, B) Filled pauses. phw = per hundred words. (C, D) False starts. (E, F) Retracings.

  • (7)
    1. and <uh they were uh> they're going to uh possibly take me <to to another uh de- uh uh uh go someplace to [sueɪɹouʔ] (=unknown)> somewhere else <to have> to get <a> uh a [biɚ] (=idea) of what was going in (=on). (JI, 3 days post, about what he did the day prior)
    2. one I did <really bad> really bad. (.) oh this one. but I get it… after a while I got it. but this one I tried. uh but then I got it. it wasn't… <but I θ-> but I <work> work on it. (JI, 5 days post, about practicing a language task)
    3. <and then uh she uh> <so we> and… but <she was> she was good. she was <in> uh I think in the hæspɪ- hospital <for about> for <about seven days I believe> seven or eight days. (JI, 7 days post, about his daughter's childhood)
    4. <and I still did a little> I still Ø (=have) [pɹɑms] (=problems) with some of the words <[tɛmp] (=that)> that won't quite [kʌmp] (=come) out <white (=right)> right. (JI, 15 days post, about his language)
    5. he said “now what you've done for the first f- three and a half wi- months it'll take a good nine or ten or eleven months more.” (JI, 110 days post, reporting his doctor's view on his recovery)
  • (8)
    1. a bunch of… (.) uh (.) and <ɪ- it's ɪ- ɪ-> <you can get (.) um (.) the (.) the the> uh you can (.) take the [tʃɜ˞mɹɑ] (=chimney) (VG, 5 days post, about building a house)
    2. and it was <on the uh h haɪ highway> (.) on a main n- highway (.) uh (.) uh uh in [tʃænəlʌ] (=Louisiana). (VG, 6 days post, about growing up on her parents' farm)
    3. anyway there's <a kæ-> the capitol <of> of ɜɪɹə- Arizona. (VG, 12 days post, about a travel memory)
    4. they were trying to get me s- to go someplace. I said “I don't wanna go anywhere. I wanna go home.” (VG, 16 days post, about people trying to move her)
    5. I've been trying to clean out that room apparently. and I just do that. and they do have people who come in and <help> help me out twice a day. (.) but it gets to the point that there's nothing else to do. (VG, 92 days post, about what she did that day)

Brief aphasia battery

All five measures improved over time in both patients (Figure 8): confrontation naming (JI: strong correlation, r = 0.96, p = 0.0045; VG: strong correlation, r = 0.93, p = 0.012), single word comprehension (JI: strong correlation, r = 0.71, p = 0.090; VG: moderate but non-significant correlation, r = 0.34, p = 0.29), sentence comprehension (JI: strong but non-significant correlation, r = 0.51, p = 0.19; VG: strong correlation, r = 0.85, p = 0.034), repetition (JI: strong correlation, r = 0.98, p = 0.0014; VG: strong correlation, r = 0.93, p = 0.010), and reading aloud (JI: strong correlation, r = 0.94, p = 0.0079; VG: strong but marginally significant correlation, r = 0.73, p = 0.082).

Figure 8.

Figure 8

Change over time in scores on the brief aphasia battery. (A, B) Confrontation naming. (C, D) Single word comprehension. (E, F) Sentence comprehension. (G, H) Repetition of words and sentences. (I, J) Reading aloud of words and sentences.

Discussion

In this study, we characterized in detail the recovery of language function over the first 3 months post-stroke in two individuals with aphasia following infarction of the left posterior superior temporal gyrus: Wernicke's area.

The first aim of our study was to determine which impaired language domains would improve over time. We found that most, but not all, language measures improved over time. Most measures associated with successful language—in particular, words per minute, mean length of utterance, number of bound morphemes, and number of embeddings—increased in both patients, while most measures of erroneous language use—in particular, phonological errors, the composite measure of impaired lexical access, and counts of filled pauses, false starts, and retracings—decreased in both patients. While not all of these changes over time were statistically significant in both patients, the data overall reflected a clear pattern of improvement across multiple language domains, including fluency, lexical access, phonological retrieval and encoding, and syntactic complexity. In VG, amelioration of dysarthria was also observed.

Not every language measure showed improvements. JI produced morphosyntactic errors at a consistent rate over time, even as the syntactic complexity of his utterances increased. Consequently, his speech at the 3-month time point remained somewhat paragrammatic. VG did not show any improvements in proportions of closed class words or verbs, which were greater than the control range, reflecting lexical access difficulties and empty speech (Wilson, Henry, et al., 2010). On the other hand, VG did show modest improvement on the lexical access composite measure.

Our second aim was to investigate the time course of recovery. We found that the language variables that changed over time were generally well fit by logarithmic models. That is, the most rapid gains were made early, and progress continued but slowed over time. By 3 months, both patients' communication skills were quite effective overall, in the sense that they were usually successful in conveying complex ideas about novel topics, as can be appreciated in some of the examples provided above. Yet both patients still had clearly apparent deficits in connected speech and on the brief aphasia battery, and both met clinical criteria for conduction aphasia, which is a typical chronic outcome for patients with posterior temporal damage (Buchsbaum et al., 2011).

The potential for amelioration of language deficits after damage to Wernicke's area demonstrated in this study is consistent with previous research (Kertesz & Benson, 1970; Kertesz et al., 1993; Kertesz & McCabe, 1977; Laska et al., 2001; Metter et al., 1990; Naeser et al., 1987; Penfield & Roberts, 1959; Selnes et al., 1983, 1984; Weiller et al., 1995), as is the rapid, decelerating time course of recovery (Basso, 1992; Pedersen et al., 1995; Swinburn, et al., 2004). The mechanisms behind this rapid recovery are still poorly understood, and may include recruitment and/or reorganization of ipsilateral perilesional tissue (Heiss, Kessler, Karbe, Fink, & Pawlik, 1993; Heiss, Kessler, Theil, Ghaemi, & Karbe, 1999; Karbe et al., 1998; Saur et al., 2006) or homotopic regions in the right hemisphere (Crinion & Price, 2005; Weiller et al., 1995). Besides the posterior STG, a broad array of temporo-parietal regions, including the SMG, MTG, angular gyrus, ITG and fusiform gyrus have been implicated in language processing in neuropsychological (Kertesz & Benson, 1970; Kertesz et al., 1993; Metter et al., 1990; Naeser et al., 1987; Selnes et al., 1983) and neuroimaging (Binder, Desai, Graves, & Conant, 2009; Binder et al., 1997) studies. The white matter fiber pathways that underlie and connect these regions, including the arcuate fasciculus, extreme capsule fiber system, inferior longitudinal fasciculus, and middle longitudinal fasciculus are also important (Catani & Mesulam, 2008; Petrides, 2014; Turken & Dronkers, 2011; Wilson et al., 2011). In the 2 patients we studied, damage was relatively circumscribed to the posterior STG and its underlying white matter, in particular the posterior and direct segments of the arcuate fasciculus (Catani & Mesulam, 2008; Turken & Dronkers, 2011). It appears that in these cases, the surviving temporo-parietal regions and white matter pathways (likely including the anterior indirect segment of the arcuate fasciculus) may provide a substrate for recovery. In contrast, lesions that extend beyond the STG, especially to the MTG, can be expected to impact a much more extensive network of white matter pathways (Turken & Dronkers, 2001), and outcomes are accordingly poorer (Kertesz & Benson, 1970; Kertesz et al., 1993; Metter et al., 1990; Naeser et al., 1987; Selnes et al., 1983).

Despite the remarkable recovery we observed, the persistent conduction aphasia we documented at 3 months post stroke contrasts with outcomes in patients with lesions that are relatively circumscribed to Broca's area, whose language has been reported to be near-normal within a month or two (Mohr, 1976). The relative persistence of language deficits after temporal damage as opposed to frontal has been reported by several other authors (Heiss et al., 1999; Penfield & Roberts, 1959; Selnes et al., 1984), and supports the view that the posterior perisylvian region is more indispensable for language than Broca's area.

Quantitative analysis of connected speech samples proved to be an effective means of tracking day-by-day recovery in the acute/subacute post-stroke period (Holland et al., 1985). Changes in language measures across time appear to reflect five main sources of variability: (1) improvement of underlying language function over time; (2) genuine fluctuations in language performance due to varying levels of arousal, motivation, attention, and so on; (3) measurement error due to the relatively brief duration of the speech samples; (4) measurement error due to subjective decisions made in transcription and coding; (5) measurement error due to quantitative measures that do not perfectly reflect underlying constructs (for example, the morphosyntactic error composite score includes non-sentence utterances, yet these are not always errors per se). While many measures showed robust logarithmic time courses, no measure was monotone increasing or decreasing, suggesting that one or more of the second, third, fourth and/or fifth factors contributed significantly to our data. We believe that the third factor—measurement error due to the relatively brief duration of the speech samples—is probably the greatest source of variability unrelated to recovery.

Tracking recovery using quantitative analysis of connected speech minimizes the impact of research on participants, which is an important consideration in the acute/subacute post-stroke period. However, while connected speech samples are relatively easy to acquire, they are extremely time consuming to transcribe and code. Transcription and coding of aphasic speech took about one hour per minute for this study. Moreover, researchers who carry out transcribing and/or coding require considerable training as well as a strong background in linguistics and/or speech-language pathology (Prins and Bastiaanse, 2004). The fact that we studied only two individuals with aphasia is the most significant limitation of our study, and reflects the time-consuming nature of our approach. To extend our approach to a larger sample of patients, personnel resources would need to be budgeted accordingly. Alternatively, it would be possible to carry out a less nuanced analysis of connected speech samples that would be less time-consuming. Two connected speech measures—words per minute (which excludes retraced words), and number of embeddings—hold particular promise as indicators of recovery, since they were strongly and significantly correlated with time post stroke in both patients. These particular measures could be derived from a less comprehensive transcription and coding scheme, and might provide measures of improvement that could have clinical or research utility. Another alternative approach that would also be less time-consuming would be to use qualitative instead of quantitative assessments of connected speech samples, as in the Boston Diagnostic Aphasia Examination (Goodglass et al., 2001), and similar to the approach that dominates the field of motor speech disorders (Darley et al., 1969). However there has been little research to date comparing qualitative and quantitative analyses of connected speech (Grande et al., 2008).

In conclusion, we documented substantial recovery of language function in the first 3 months post stroke in two individuals with aphasia following infarction of the left posterior superior temporal gyrus. Recovery was apparent across multiple language domains, including fluency, lexical access, phonological retrieval and encoding, and syntactic complexity. Most language measures improved with a logarithmic time course, with the greatest gains taking place in early in the course of recovery.

Acknowledgments

We thank Audrey Holland for her advice on collecting and analyzing speech samples; Pélagie Beeson and Gayle DeDe for providing input on project design and constructive feedback on a draft of this manuscript; Grace Cheifetz, Jane Eustance, Lisa Gordon, Sara McDonald, and Kim Neely for rating motor speech characteristics of speech samples; Kindle Rising and Dana Eriksson for helpful discussions; three anonymous reviewers for their constructive feedback; and the patients for their participation in our research.

Funding: This research was supported in part by the National Institutes of Health (National Institute on Deafness and Other Communication Disorders) under grant R01 DC013270.

References

  1. Azuar C, Leger A, Arbizu C, Henry-Amar F, Chomel-Guillaume S, Samson Y. The aphasia rapid test: an NIHSS-like aphasia test. Journal of Neurology. 2013;260:2110–2117. doi: 10.1007/s00415-013-6943-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  2. Basso A. Prognostic factors in aphasia. Aphasiology. 1992;6:337–348. [Google Scholar]
  3. Behrens TEJ, Berg HJ, Jbabdi S, Rushworth MFS, Woolrich MW. Probabilistic diffusion tractography with multiple fibre orientations: What can we gain? NeuroImage. 2007;34:144–155. doi: 10.1016/j.neuroimage.2006.09.018. [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Benjamini Y, Hochberg Y. Controlling the false discovery rate: A practical and powerful approach to multiple testing. Journal of the Royal Statistical Society Series B. 1995;57:289–300. [Google Scholar]
  5. Binder JR, Desai RH, Graves WW, Conant LL. Where is the semantic system? A critical review and meta-analysis of 120 functional neuroimaging studies. Cerebral Cortex. 2009;19:2767–2796. doi: 10.1093/cercor/bhp055. [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Binder JR, Frost JA, Hammeke TA, Cox RW, Rao SM, Prieto T. Human brain language areas identified by functional magnetic resonance imaging. Journal of Neuroscience. 1997;17:353–362. doi: 10.1523/JNEUROSCI.17-01-00353.1997. [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Bogen JE, Bogen GM. Wernicke's region—Where is it? Annals of the New York Academy of Sciences. 1976;280:834–843. doi: 10.1111/j.1749-6632.1976.tb25546.x. [DOI] [PubMed] [Google Scholar]
  8. Buchsbaum BR, Baldo J, Okada K, Berman KF, Dronkers N, D'Esposito M, Hickok G. Conduction aphasia, sensory-motor integration, and phonological short-term memory – An aggregate analysis of lesion and fMRI data. Brain and Language. 2011;119:119–128. doi: 10.1016/j.bandl.2010.12.001. [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Catani M, Mesulam M. The arcuate fasciculus and the disconnection theme in language and aphasia: history and current state. Cortex. 2008;44:953–961. doi: 10.1016/j.cortex.2008.04.002. [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Cohen J. Statistical Power Analysis for the Behavioral Sciences. 2nd. Hillsdale, NJ: Lawrence Erlbaum; 1988. [Google Scholar]
  11. Crinion J, Price CJ. Right anterior superior temporal activation predicts auditory sentence comprehension following aphasic stroke. Brain. 2005;128:2858–2871. doi: 10.1093/brain/awh659. [DOI] [PubMed] [Google Scholar]
  12. Darley FL, Aronson AE, Brown JR. Differential diagnostic patterns of dysarthria. Journal of Speech and Hearing Research. 1969;12:246–269. doi: 10.1044/jshr.1202.246. [DOI] [PubMed] [Google Scholar]
  13. Duffy JR. Motor speech disorders: Substrates, differential diagnosis, and management. 3rd. St. Louis: Elsevier/Mosby; 2013. [Google Scholar]
  14. Flamand-Roze C, Falissard B, Roze E, Maintigneux L, Beziz J, Chacon A, et al. Denier C. Validation of a new language screening tool for patients with acute stroke: the Language Screening Test (LAST) Stroke. 2011;42:1224–1229. doi: 10.1161/STROKEAHA.110.609503. [DOI] [PubMed] [Google Scholar]
  15. Folstein MF, Folstein SE, McHugh PR. “Mini-mental state” : A practical method for grading the cognitive state of patients for the clinician. Journal of Psychiatric Research. 1975;12:189–198. doi: 10.1016/0022-3956(75)90026-6. [DOI] [PubMed] [Google Scholar]
  16. Friston KJ. Statistical parametric mapping: The analysis of funtional brain images. Amsterdam; Boston: Elsevier/Academic Press; 2007. [Google Scholar]
  17. Galantucci S, Tartaglia MC, Wilson SM, Henry ML, Filippi M, Agosta F, et al. Gorno-Tempini ML. White matter damage in primary progressive aphasias: A diffusion tensor tractography study. Brain. 2011;134:3011–3029. doi: 10.1093/brain/awr099. [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Gandour J, Marshall R, Kim S, Neuburger S. On the nature of conduction aphasia: A longitudinal case study. Aphasiology. 1991;5:291–306. [Google Scholar]
  19. Goodglass H, Kaplan E, Barresi B. The Boston Diagnostic Aphasia Examination (BDAE) 3rd. Baltimore: Lippincott Williamson & Wilkins; 2001. [Google Scholar]
  20. Grande M, Hussmann K, Bay E, Christoph S, Piefke M, Willmes K, Huber W. Basic parameters of spontaneous speech as a sensitive method for measuring change during the course of aphasia. International Journal of Language & Communication Disorders. 2008;43:408–426. doi: 10.1080/13682820701685991. [DOI] [PubMed] [Google Scholar]
  21. Heiss WD, Kessler J, Karbe H, Fink GR, Pawlik G. Cerebral glucose metabolism as a predictor of recovery from aphasia in ischemic stroke. Archives of Neurology. 1993;50:958–964. doi: 10.1001/archneur.1993.00540090059011. [DOI] [PubMed] [Google Scholar]
  22. Heiss WD, Kessler J, Thiel A, Ghaemi M, Karbe H. Differential capacity of left and right hemispheric areas for compensation of poststroke aphasia. Annals of Neurology. 1999;45:430–438. doi: 10.1002/1531-8249(199904)45:4<430::aid-ana3>3.0.co;2-p. [DOI] [PubMed] [Google Scholar]
  23. Hillis AE, Wityk RJ, Tuffiash E, Beauchamp NJ, Jacobs MA, Barker PB, Selnes OA. Hypoperfusion of Wernicke's area predicts severity of semantic deficit in acute stroke. Annals of Neurology. 2001;50:561–566. doi: 10.1002/ana.1265. [DOI] [PubMed] [Google Scholar]
  24. Holland AL, Miller J, Reinmuth OM, Bartlett C, Fromm D, Pashek G, et al. Swindell C. Rapid recovery from aphasia: A detailed language analysis. Brain and Language. 1985;24:156–173. doi: 10.1016/0093-934x(85)90101-4. [DOI] [PubMed] [Google Scholar]
  25. Karbe H, Thiel A, Weber-Luxenburger G, Herholz K, Kessler J, Heiss WD. Brain plasticity in poststroke aphasia: What is the contribution of the right hemisphere? Brain and Language. 1998;64:215–230. doi: 10.1006/brln.1998.1961. [DOI] [PubMed] [Google Scholar]
  26. Kertesz A, Benson DF. Neologistic jargon: a clinicopathological study. Cortex. 1970;6:362–386. doi: 10.1016/s0010-9452(70)80002-8. [DOI] [PubMed] [Google Scholar]
  27. Kertesz A, Lau WK, Polk M. The Structural determinants of recovery in Wernicke's aphasia. Brain and Language. 1993;44:153–164. doi: 10.1006/brln.1993.1010. [DOI] [PubMed] [Google Scholar]
  28. Kertesz A, McCabe P. Recovery patterns and prognosis in aphasia. Brain. 1977;100:1–18. doi: 10.1093/brain/100.1.1. [DOI] [PubMed] [Google Scholar]
  29. Kohn S, Smith KL. Evolution of impaired access to the phonological lexicon. Journal of Neurolinguistics. 1994;8:267–288. [Google Scholar]
  30. Laska AC, Hellblom A, Murray V, Kahan T, Von Arbin M. Aphasia in acute stroke and relation to outcome. Journal of Internal Medicine. 2001;249:413–422. doi: 10.1046/j.1365-2796.2001.00812.x. [DOI] [PubMed] [Google Scholar]
  31. MacWhinney B. The CHILDES Project: Tools for Analyzing Talk. Mahwah, NJ: Lawrence Erlbaum; 2000. [Google Scholar]
  32. MacWhinney B. The CHILDES Project: Tools for Analyzing Talk—Electronic Edition. MS: Carnegie Mellon University; 2012. Aug 6, 2012. [Google Scholar]
  33. MacWhinney B, Fromm D, Forbes M, Holland A. AphasiaBank: Methods for Studying Discourse. Aphasiology. 2011;25:1286–1307. doi: 10.1080/02687038.2011.589893. [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Marie P. The third frontal convolution plays no special role in the function of language. Semaine Medicale. 1906;26:241–247. [Google Scholar]
  35. Marshall RC. Language and speech recovery in a case of viral encephalitis. Brain and Language. 1982;17:316–326. doi: 10.1016/0093-934x(82)90024-4. [DOI] [PubMed] [Google Scholar]
  36. Metter EJ, Hanson WR, Jackson CA, Kempler D, van Lancker D, Mazziotta JC, Phelps ME. Temporoparietal cortex in aphasia. Evidence from positron emission tomography. Archives of Neurology. 1990;47:1235–1238. doi: 10.1001/archneur.1990.00530110097024. [DOI] [PubMed] [Google Scholar]
  37. Mohr JP. Broca's area and Broca's aphasia. In: Whitaker H, editor. Studies in Neurolinguistics. New York: Academic Press; 1976. pp. 201–233. [Google Scholar]
  38. Naeser MA, Helm-Estabrooks N, Haas G, Auerbach S, Srinivasan M. Relationship between lesion extent in ‘Wernicke's area’ on computed tomographic scan and predicting recovery of comprehension in Wernicke's aphasia. Archives of Neurology. 1987;44:73–82. doi: 10.1001/archneur.1987.00520130057018. [DOI] [PubMed] [Google Scholar]
  39. Pedersen PM, Stig Jørgensen H, Nakayama H, Raaschou HO, Olsen TS. Aphasia in acute stroke: Incidence, determinants, and recovery. Annals of Neurology. 1995;38:659–666. doi: 10.1002/ana.410380416. [DOI] [PubMed] [Google Scholar]
  40. Penfield W, Roberts L. Speech and Brain Mechanisms. Princeton, New Jersey: Princeton University Press; 1959. [Google Scholar]
  41. Petrides M. Neuroanatomy of language regions of the human brain. Amsterdam: Elsevier; 2014. [Google Scholar]
  42. Prins R, Bastiaanse R. Analysing the spontaneous speech of aphasic speakers. Aphasiology. 2004;18:1075–1091. [Google Scholar]
  43. Rossion B, Pourtois G. Revisiting Snodgrass and Vanderwart's object pictorial set: The role of surface detail in basic-level object recognition. Perception. 2004;33:217–36. doi: 10.1068/p5117. [DOI] [PubMed] [Google Scholar]
  44. Saur D, Lange R, Baumgaertner A, Schraknepper V, Willmes K, Rijntjes M, Weiller C. Dynamics of language reorganization after stroke. Brain. 2006;129:1371–1384. doi: 10.1093/brain/awl090. [DOI] [PubMed] [Google Scholar]
  45. Saffran EM, Berndt RS, Schwartz MF. The quantitative analysis of agrammatic production: procedure and data. Brain and Language. 1989;37:440–479. doi: 10.1016/0093-934x(89)90030-8. [DOI] [PubMed] [Google Scholar]
  46. Selnes OA, Knopman DS, Niccum N, Rubens AB, Larson D. Computed tomographic scan correlates of auditory comprehension deficits in aphasia: A prospective recovery study. Annals of Neurology. 1983;13:558–566. doi: 10.1002/ana.410130515. [DOI] [PubMed] [Google Scholar]
  47. Selnes OA, Niccum N, Knopman DS, Rubens AB. Recovery of single word comprehension: CT-scan correlates. Brain and Language. 1984;21:72–84. doi: 10.1016/0093-934x(84)90037-3. [DOI] [PubMed] [Google Scholar]
  48. Swinburn K, Porter G, Howard D. Comprehensive Aphasia Test. Hove: Psychology Press; 2004. [Google Scholar]
  49. Turken AU, Dronkers NF. The neural architecture of the language comprehension network: converging evidence from lesion and connectivity analyses. Frontiers in Systems Neuroscience. 2011;5:1. doi: 10.3389/fnsys.2011.00001. [DOI] [PMC free article] [PubMed] [Google Scholar]
  50. Weiller C, Isensee C, Rijntjes M, Huber W, Müller S, Bier D, et al. Diener HC. Recovery from Wernicke's aphasia: a positron emission tomographic study. Annals of Neurology. 1995;37:723–732. doi: 10.1002/ana.410370605. [DOI] [PubMed] [Google Scholar]
  51. Wernicke K. Der Aphasische Symptomencomplex. Breslau: Cohn and Weigert; 1874. [Google Scholar]
  52. Wilson SM, Dronkers NF, Ogar JM, Jang J, Growdon ME, Agosta F, et al. Gorno-Tempini ML. Neural correlates of syntactic processing in the nonfluent variant of primary progressive aphasia. Journal of Neuroscience. 2010;30:16845–16854. doi: 10.1523/JNEUROSCI.2547-10.2010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  53. Wilson SM, Galantucci S, Tartaglia MC, Rising K, Patterson DK, Henry ML, et al. Gorno-Tempini ML. Syntactic processing depends on dorsal language tracts. Neuron. 2011;72:397–403. doi: 10.1016/j.neuron.2011.09.014. [DOI] [PMC free article] [PubMed] [Google Scholar]
  54. Wilson SM, Henry ML, Besbris M, Ogar JM, Dronkers NF, Jarrold W, et al. Gorno-Tempini ML. Connected speech production in three variants of primary progressive aphasia. Brain. 2010;133:2069–2088. doi: 10.1093/brain/awq129. [DOI] [PMC free article] [PubMed] [Google Scholar]
  55. Wittenburg P, Brugman H, Russel A, Klassmann A, Sloetjes H. ELAN: A professional framework for multimodality research; Proceedings of LREC 2006, Fifth International Conference on Language Resources and Evaluation.2006. [Google Scholar]
  56. Yushkevich PA, Piven J, Hazlett HC, Smith RG, Ho S, Gee JC, Gerig G. User-guided 3D active contour segmentation of anatomical structures: significantly improved efficiency and reliability. NeuroImage. 2006;31:1116–1128. doi: 10.1016/j.neuroimage.2006.01.015. [DOI] [PubMed] [Google Scholar]

RESOURCES