Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2015 Aug 1.
Published in final edited form as: J Speech Lang Hear Res. 2014 Aug;57(4):1480–1502. doi: 10.1044/2014_JSLHR-L-13-0192

Children with a history of SLI show reduced sensitivity to audiovisual temporal asynchrony: An ERP Study

Natalya Kaganovich 1,2,*, Jennifer Schumaker 1, Laurence B Leonard 1, Dana Gustafson 1, Danielle Macias 1
PMCID: PMC4266431  NIHMSID: NIHMS646425  PMID: 24686922

Abstract

Purpose

We examined whether school-age children with a history of SLI (H-SLI), their typically developing (TD) peers, and adults differ in sensitivity to audiovisual temporal asynchrony and whether such difference stems from the sensory encoding of audiovisual information.

Method

15 H-SLI children, 15 TD children, and 15 adults judged whether a flashed explosion-shaped figure and a 2 kHz pure tone occurred simultaneously. The stimuli were presented at 0, 100, 200, 300, 400, and 500 ms temporal offsets. This task was combined with EEG recordings.

Results

H-SLI children were profoundly less sensitive to temporal separations between auditory and visual modalities compared to their TD peers. Those H-SLI children who performed better at simultaneity judgment also had higher language aptitude. TD children were less accurate than adults, revealing a remarkably prolonged developmental course of the audiovisual temporal discrimination. Analysis of early ERP components suggested that poor sensory encoding was not a key factor in H-SLI children’s reduced sensitivity to audiovisual asynchrony.

Conclusions

Audiovisual temporal discrimination is impaired in H-SLI children and is still immature during mid-childhood in TD children. The present findings highlight the need for further evaluation of the role of atypical audiovisual processing in the development of SLI.

1. Introduction

A growing body of literature suggests that visual speech cues play an important role in typical language acquisition and speech perception. For example, the presence of a talker’s face has been shown to enhance infants’ ability to learn phoneme boundaries (Teinonen, Aslin, Alku, & Csibra, 2008) and new words (Hollich, Newman, & Jusczyk, 2005). It also significantly improves the understanding of speech-in-noise in both children and adults (Barutchu et al., 2010; Sumby & Pollack, 1954). The facilitative effect of visual cues on auditory perception is especially strong in individuals with hearing impairment. Thus, prelingually deaf children with cochlear implants understand speech better when it is audiovisual as compared to auditory only (Bergeson, Pisoni, & Davis, 2005). The degree of such benefit depends on the age of implantation, underlining the significance of audiovisual stimulation during early childhood (Schorr, Fox, van Wassenhove, & Knudsen, 2005).

The importance of visual speech cues during language acquisition is further bolstered by the findings that infants learn to pay attention to the mouth of the speaker during the first year of life. For example, 4–8 month old infants fixate on the mouth, rather than on the eyes, of a talker, presumably to benefit from redundant visual cues while learning the phonemes of their native language (Lewkowicz & Hansen-Tift, 2012). In fact, infants appear to be so acutely sensitive to articulatory movements that they are able to tell two different languages apart based on visual cues alone (Weikum et al., 2007). Paying attention to visual speech cues continues to be important during adulthood. A recent study by Yi and colleagues (Yi, Wong, & Eizenman, 2013) showed that as speech-to-noise ratio decreases, adults fixate their gaze significantly more on the speaker’s mouth in order to maintain good speech perception.

Several behavioral and electrophysiological studies have employed the so-called McGurk illusion in order to investigate how early in development visual information can influence auditory speech perception (Burnham & Dodd, 2004; Kushnerenko, Teinonen, Volein, & Csibra, 2008; Rosenblum, Schmuckler, & Johnson, 1997). The McGurk illusion is perceived when visual and auditory cues to phoneme identity are in conflict with each other. Although the exact stimuli used vary from study to study, it is common to dub the auditory “pa” onto the visual articulation of “ka,” with the resulting perception being that of “ta” if the merger of auditory and visual cues takes place. Such studies have demonstrated that infants as young as 4.5–5 months of age are in fact influenced by visual information and perceive the McGurk illusion. A study by Bristow and colleagues, using a different paradigm, examined audiovisual processing in even younger infants – just 10 weeks of age – and showed that infants integrate speech-related auditory and visual information during early stages of perception, similar to adults (Bristow et al., 2008).

Although a substantial amount of audiovisual processing is present early in development, as is evidenced by the above studies, the ability to fully integrate different senses and thus to gain the most benefit from multisensory information may have a protracted developmental course. For example, multiple studies have reported a reduced susceptibility to the McGurk illusion in children compared to adults and suggested that children are overall less influenced by visual information during auditory perception (e.g., Massaro, 1984; Massaro, Thompson, Barron, & Lauren, 1986; McGurk & MacDonald, 1976). In agreement with these behavioral reports, a neuroimaging study by Dick and colleagues has shown that the functional connectivity between brain regions involved in audiovisual speech perception differs significantly between school-age children and adults, indicating that audiovisual integration may not yet be mature by mid-childhood (Dick, Solodkin, & Small, 2010). This conclusion is also supported by an electrophysiological study of Brandwein and colleagues who reported that multisensory facilitation on a simple reaction time task and the accompanying reduction in the amplitude of the N1 ERP component to multisensory stimuli does not reach adult levels until approximately 14 years of age (Brandwein et al., 2011).

Despite the accumulating evidence for the importance of audiovisual processing in language acquisition, very little is known about a relationship between audiovisual processing and language aptitude. If visual speech cues facilitate language acquisition, then we would expect that those children who are good audiovisual integrators should also be effective language learners. On the other hand, a breakdown in efficient audiovisual processing may be expected to delay or disrupt at least some aspects of language acquisition and thus, hypothetically, contribute to the development of language disorders. A few studies have examined this issue by focusing on children with specific language impairment (SLI) (Boliek, Keintz, Norrix, & Obrzut, 2010; Hayes, Tiippana, Nicol, Sams, & Kraus, 2003; Meronen, Tiippana, Westerholm, & Ahonen, 2013; Norrix, Plante, & Vance, 2006; Norrix, Plante, Vance, & Boliek, 2007). SLI is a language disorder that is associated with significant linguistic difficulties in the absence of hearing impairment, frank neurological disorders, or low non-verbal intelligence (Leonard, 2014). Although it is typically diagnosed in early childhood, many school-age children with a history of SLI who no longer make overt grammatical errors continue to lag behind their peers in language aptitude (e.g., Hesketh & Conti-Ramsden, 2013), revealing the long-term nature of this disorder.

In all studies of audiovisual processing in SLI to date, the McGurk paradigm was used, and the authors reported the amount of fused McGurk perceptions and visual captures (e.g, hearing “ka” when the stimulus consisted of a visual “ka” combined with an auditory “pa”) in SLI children compared to their TD peers. Although as a whole these studies point to significant differences in the processing of audiovisual speech between SLI and TD children, specific findings and their interpretations vary from study to study. Most studies report a reduced number of fused McGurk percepts in children with language impairment (e.g., Boliek et al., 2010; Hayes et al., 2003; Norrix et al., 2007), with one exception being the study by Meronen and colleagues (Meronen et al., 2013). Findings on the amount of visual capture also differ. Hayes and colleagues reported that compared to TD children, children with language-learning difficulties showed significantly more instances of visual capture at low signal-to-noise ratios, which suggested that they relied primarily on visual information when responding (Hayes et al., 2003). On the other hand, a recent study by Meronen and colleagues showed the opposite effect, with a lower rate of visual capture in SLI compared to TD children under most difficult listening conditions, concluding that children with SLI have weak lip-reading skills (Meronen et al., 2013).

1.1. Audiovisual Temporal Discrimination

The reason for differences in the processing of audiovisual speech between SLI and TD children remains unclear. One possibility is that the McGurk stimulus, with its conflicting auditory and visual information about the identity of the phoneme, may present a challenge to the SLI population whose phonological awareness is thought to be weaker than that of TD children (e.g., Catts, 1993). However, the problem may also stem from difficulty in detecting temporal correspondences between auditory and visual information. Extensive research in the area of audiovisual integration shows that a temporal alignment between different sensory modalities is one of the key factors determining a successful merger of the senses (e.g., Meredith, Nemitz, & Stein, 1987). A reduced sensitivity to temporal discrepancies in audiovisual stimuli might, therefore, be a factor in poor audiovisual integration in SLI.

Individuals perceive audiovisual stimuli as synchronous not only when the two modalities are perfectly aligned in time, but also when they are separated by relatively short temporal periods of several tens to several hundreds of milliseconds (Bushara, Grafman, & Hallett, 2001; Dixon & Spitz, 1980; Grant, van Wassenhove, & Poeppel, 2004; McGrath & Summerfield, 1985; Petrini, Dahl, et al., 2009; Petrini, Halt, & Pollick, 2010; Petrini, Russell, & Pollick, 2009; Van Wassenhove, Grant, & Poeppel, 2007; Vatakis & Spence, 2006; for a review, see also Vroomen & Keetels, 2010; Vroomen & Stekelenburg, 2009; Vroomen & Stekelenburg, 2011). This temporal window of integration is significantly larger in infants but shortens with development (Lewkowicz, 1996, 2003, 2010; Pons, Lewkowicz, Soto-Faraco, & Sebastián-Gallés, 2009). However, the majority of studies on temporal audiovisual processing have been conducted with adults, and its maturational course is yet to be fully mapped. In two recent pioneering studies, Hillock-Dunn and colleagues examined audiovisual temporal discrimination in several groups of school-age children and in adults using a simultaneity judgment task (Hillock-Dunn & Wallace, 2012; Hillock, Powers, & Wallace, 2011). They reported that both children (6–11 years of age) and adolescents (12–17 years of age) were significantly more likely than adults to identify audiovisual stimuli separated by medium to large offsets as synchronous, pointing to a protracted developmental course of the audiovisual temporal function.

Temporal perception in children with SLI has been studied in great detail, and a deficit in the temporal discrimination mechanism has even been suggested as one of the main causes of this disorder (e.g., Tallal, Stark, & Mellits, 1985). However, almost all studies in this area used a single modality (mostly auditory) and focused primarily on the effect of stimulus brevity and the speed of presentation on auditory perception. To the best of our knowledge, only a few studies examined multimodal temporal discrimination in SLI. Tallal and colleagues (Tallal et al., 1985) reported that performance on sequencing cross-modal non-verbal stimuli with short inter-stimulus intervals (ISI) was one of six variables that, when taken together, correctly identified 98% of children with SLI in their sample as language-impaired. Short ISIs ranged from 8 ms to 305 ms (Tallal, 1980), and whether all were equally predictive of group membership is unknown. Because in Tallal’s studies children were asked to reproduce the order in which stimuli appeared, it is difficult to determine whether children were unable to perceive the correct sequence of stimuli or unable to maintain it in their working memory. A study by Grondin and colleagues (Grondin et al., 2007) tested 6–9 year old children with SLI and their age-matched TD peers on two temporal order judgment tests. In one, children had to report which of two flashes, on the left or on the right, appeared first. In another, children indicated whether the flash or the tone appeared first. The authors found that children with SLI performed significantly worse than their TD peers on both tasks; however, they did especially poorly on the bimodal task. Here, the SLI group’s accuracy remained close to chance level, even when separations between the onsets of the flash and the tone were as large as 420 ms. However, the stimuli used by Grondin and colleagues were only 5 ms in duration. Because children with SLI are known to have difficulty with very short stimuli (Leonard, McGregor, & Allen, 1992), their performance might have reflected the struggle to efficiently process brief sounds and flashes, rather than a failure to determine their temporal alignment.

Lastly, a recent study by Pons and colleagues (Pons, Andreu, Sanz-Torrent, Buil-Legaz, & Lewkowicz, 2013) compared the ability of 4–7 year old Spanish-speaking children with and without SLI to detect audiovisual temporal asynchrony in speech by evaluating how long children fixated on each of two simultaneously presented video clips. In one video clip, the audio and video portions were always synchronized, while in another they were offset by either 366 ms or 666 ms and presented with either the audio leading the video (the AV condition) or the video leading the audio (the VA condition). The authors assumed that if children can tell the difference between synchronous and asynchronous videos, they would prefer to fixate on the synchronous one. They found that neither group of children was able to detect speech asynchrony at the 366 ms offset. Further, while TD children detected the asynchrony at the 666 ms offset regardless of whether the audio preceded or followed the video, the SLI children detected temporal asynchrony only in the AV condition thus displaying a reduced ability to perceive audiovisual temporal asynchrony in speech.

1.2 Current Study

Taken together, the above studies suggest that children with language learning difficulties may in fact be impaired at detecting temporal correspondences between auditory and visual information; however, a more definitive evaluation with stimuli that do not place significant demands on either sensory or cognitive processing is needed. Additionally, the reason behind potentially worse audiovisual temporal discrimination in SLI remains elusive. Such measures as reaction time and accuracy represent the end result of multiple sensory and cognitive stages of processing and thus cannot identify the mechanism or mechanisms responsible for atypical audivovisual temporal function in SLI. One possibility is that children with SLI are impaired at detecting onsets of auditory and visual stimuli and encoding their early sensory properties when such stimuli occur in close temporal proximity. However, the difficulty may also stem from later cognitive processes, such as making an overt judgment about the temporal coincidence of two modalities. In order to understand the nature of neural processes that differ between TD and SLI children during audiovisual temporal discrimination, we combined a simultaneity judgment task (SJT) with event-related potentials (ERPs) recordings. ERPs are produced by large assemblies of post-synaptic neurons that fire in synchrony in response to a stimulus and are recorded on the scalp surface. This technique has exquisite temporal resolution, which allows one to track brain responses on a millisecond (ms) scale. Importantly, ERPs recorded within approximately 200 ms post-stimulus onset in either auditory or visual modalities are thought to reflect primarily sensory encoding of a stimulus and thus allow one to examine sensory processing somewhat independently from later cognitive analyses. This early temporal window typically includes a positive deflection called P1 and a negative deflection called N1.

In the visual modality, the P1 and N1 components are sensitive to physical properties of visual stimuli, such as luminance or pattern change. Infant studies show that the visual P1 to flash stimuli appears during the first 15 weeks of life (McCulloch, 2007). Its latency slowly shortens and the amplitude decreases during development. The early part of P1 is thought to originate in the dorsal extrastriate cortex of the middle occipital gyrus while the latter part from the ventral extrastriate cortex of the fusiform gyrus (Di Russo, Martínez, Sereno, Pitzalis, & Hillyard, 2001). The visual P1 component is followed by N1. Surface-recorded visual N1 is thought to be the result of activity in the lateral extrastriate cortex and inferior occipital-temporal cortex (Pratt, 2011). The visual P1 and N1 components have an occipital distribution in an ERP waveform.

In the auditory modality, the P1 and N1 components are thought to reflect the early sensory encoding of physical properties of sound, such as frequency, complexity, and intensity, among others (Näätänen & Picton, 1987). P1 is one of the dominant features of children’s ERPs, and its amplitude diminishes with age (Čeponienė, Rinne, & Näätänen, 2002). P1 is uniquely sensitive to sound onsets and is often absent for sound offsets. This component is likely the result of multiple neural generators, which include the primary auditory cortex, but also the hippocampus, the planum temporale, and the lateral temporal cortex (Pratt, 2011). The auditory N1 appears in children’s ERP recordings during the first few years of life. Developmental studies of early auditory ERPs sometimes disagree on the nature of the ERP components present during middle childhood, with some studies reporting immature ERPs in this age group (e.g., Sharma, Kraus, McGee, & Nicol, 1997; Sussman, Steinschneider, Gumenyuk, Grushko, & Lawson, 2008) while others showing an adult-like sequence of P1, N1, and P2 components (e.g., Čeponienė et al., 2002; Pang & Taylor, 2000). At least some of such disagreements are likely due to differences in the rate of stimulus presentation in various studies, with more adult-like ERPs elicited by stimuli presented at longer inter-stimulus intervals. Näätänen and Picton (1987) provide evidence for multiple N1 sub-components, which, together, may contribute to its parameters in the average waveform. At least one of these subcomponents is believed to originate from the supratemporal plane of the brain (Näätänen & Picton, 1987; Scherg, Vajsar, & Picton, 1989), which houses the primary auditory cortex. The auditory P1 and N1 components have fronto-central distribution in an ERP waveform.

Previous research suggests that some of these early auditory and visual ERP components may be atypical at least in a subset of children with SLI. The majority of studies on early processing in SLI focused on the auditory modality and examined the N1 peak amplitude and latency to either brief sounds or sounds that were presented in quick succession. For example, Neville and colleagues (Neville, Coffey, Holcomb, & Tallal, 1993) reported that N1 to standard tones was reduced at short ISIs over the right hemisphere sites in school-age language-impaired children. A similar finding was also reported for adolescents with SLI (Weber-Fox, Leonard, Hampton Wray, & Tomblin, 2010). Similarly, McArthur and Bishop (Bishop & McArthur, 2004; McArthur & Bishop, 2005) showed that children and young adults with SLI have age-inappropriate ERP components (including N1) to a range of sounds, not just pure tones. Stevens and colleagues suggested that the auditory N1 reduction in children with SLI is due to the attenuated auditory recovery cycle (Stevens, Paulsen, Yasen, Mitsunaga, & Neville, 2012). Studies that specifically examined visual ERPs in children with SLI are very few. In a seminal paper, Neville and colleagues (Neville et al., 1993) reported a reduction in the amplitude of visual P1 in school-age children with language impairment. However, Neville and colleagues’ participants had severe reading difficulties indicative of developmental dyslexia; therefore, whether this finding would extend to children with SLI without dyslexia is unclear. In sum, although a number of studies examined early sensory processing of auditory and visual modalities in SLI, most of such studies looked at one modality at a time. As a result, little is known about sensory processing in SLI under cross-modal conditions.

In this study, we compared the performance of school-age children with a history of SLI, their TD peers, and adults on a simultaneity judgment task (SJT) that was presented as a game and contained stimuli that do not pose sensory or cognitive processing difficulty. In order to avoid the influence of language skills on the SJT performance, we used non-speech stimuli – a pure tone and a flash of light – whose pairing was made meaningful through the context of the game. These stimuli were presented either synchronously or separated by temporal intervals that ranged from 100 to 500 ms, in 100 ms increments. The goal of behavioral analysis was to compare the accuracy with which children and adults could detect temporal asynchrony as offsets between modalities increased. The goal of our ERP analyses was to determine whether children with language learning difficulties differed in early neural response to auditory and visual stimuli as a function of the temporal separation between the two modalities. If found, such a difference would point to sensory encoding as a contributing factor to audiovisual temporal discrimination difficulty in SLI. Its absence, on the other hand, would suggest that group differences likely lie in later, cognitive stages of processing of temporal relationships. Lastly, in order to better understand the role of multisensory temporal discrimination in language skills, we performed linear regressions between children’s performance on the SJT and their language aptitude as measured by standardized tests.

2. Method

2.1 Participants

Three separate groups of participants took part in the study: 15 children with a history of SLI (H-SLI) (3 female), 15 typically developing (TD) children (4 female), and 15 college-age adults (4 female). The average age of the H-SLI group was 9 years (range: 7 years 3 months to 11 years 0 months), the average age of the TD group was 9 years and 1 month (range: 7 years 4 months to 10 years 10 months), and the average age of the adult group was 23 years (range: 20–40 years). The two groups of children were age-matched within approximately 3 months. All participants were monolingual native speakers of American English. All gave their written consent or assent to participate in the experiment. Additionally, at least one parent of each child gave a written consent to enroll their child in the study. The study was approved by the Institutional Review Board of Purdue University, and all study procedures conformed to The Code of Ethics of the World Medical Association (Declaration of Helsinki) (1964).

All children were administered the following battery of tests. Language skills were evaluated with the Clinical Evaluation of Language Fundamentals – 4th edition (CELF-4; Semel, Wiig, & Secord, 2003). Each child was tested on the four sub-tests that, together, provide the Core Language Score (CLS) as an overall measure of the child’s language ability. These sub-tests differed depending on the children’s age. All children were administered the Concepts and Following Directions, Recalling Sentences, and Formulated Sentences sub-tests. Additionally, 7–8 year olds (H-SLI n=8, TD n=7) were also tested on the Word Structure (WS) sub-test while 9–11 year olds (H-SLI n=7, TD n=8) were tested on the Word Classes-2 Total (WC) sub-test. Children’s verbal working memory was evaluated with the non-word repetition test (Dollaghan & Campbell, 1998) and the Number Memory Forward and Number Memory Reversed sub-tests of the Test of Auditory Processing Skills – 3rd edition (TAPS-3; Martin & Brownel, 2005). Non-verbal intelligence was assessed with the Test of Nonverbal Intelligence - 4th edition (TONI-4; Brown, Sherbenou, & Johnsen, 2010). Because the main task of the study required sustained attention and the ability to pay attention to two modalities presented at the same (or nearly the same) time, we examined children’s attention skills by administering the following sub-tests of the Test of Everyday Attention for Children (TEA-Ch; Manly, Robertson, Anderson, & Nimmo-Smith, 1999): Sky Search, Sky Search Dual Task, and Map Mission. The first two sub-tests required that children find and circle matching pairs of space ships surrounded by similar-looking distractor ships. During the Dual Task version of Sky Search, children also counted concurrently presented tone pips. Lastly, in Map Mission, children were asked to circle as many “knives-and-forks” symbols on a city map as possible during 1 minute. These tasks measured children’s ability to selectively focus or divide their attention.

The presence of Attention Deficit/Hyperactivity Disorder (ADHD) was evaluated with the help of the short version of the Parent Rating Scale of the Conners’ Rating Scales – Revised (Conners, 1997) and via a medical history form filled out by children’s parents. To rule out the presence of autism spectrum disorders, all children were evaluated with the Childhood Autism Rating Scale - 2nd edition (CARS-2; Schopler, Van Bourgondien, Wellman, & Love, 2010). We also recorded the level of mothers’ and fathers’ education as a measure of children’s socioeconomic status (SES). Data on adult participants’ non-verbal intelligence were collected with the help of the TONI-4. In all participants, handedness was assessed with an augmented version of the Edinburgh Handedness Questionnaire (Cohen, 2008; Oldfield, 1971).

Children in the H-SLI group were classified as having SLI when they were 4–5 years of age. These children were described in several studies of Leonard and colleagues (Leonard, Davis, & Deevy, 2007; Leonard & Deevy, 2011; Leonard, Deevy, et al., 2007). The gold standard for their evaluation at that time consisted of either the Structured Photographic Expressive Language Test – 2nd Edition (SPELT - II, Werner & Kresheck, 1983) or the Structured Photographic Expressive Language Test – Preschool 2 (Dawson, Eyer, & Fonkalsrud, 2005). Both tests have shown good sensitivity and specificity (Greenslade, Plante, & Vance, 2009; Plante & Vance, 1994). During diagnosis, all children fell below the 10th percentile on the above tests, with the majority falling below the 2nd percentile (range 1st – 8th percentile). During the years between the original diagnosis and participating in the current study, all but two H-SLI children received some form of speech therapy, which ranged in length from 2 to 8 years. Although at the time of this study, as a group, the H-SLI children scored within the normal range of language abilities based on the CLS of CELF-4, their CLS scores were nonetheless significantly lower than those of the TD group (F(1,29)=9.427, p=0.005) (see Table 1 for individual data and group means for CELF-4 subtests and CLS). Out of the four sub-tests of the CELF-4 that make up the composite CLS, the H-SLI group performed significantly worse on two – Formulated Sentences (F(1,29)=5.077, p=0.032) and Recalling Sentences (F(1,29)=16.413, p<0.001). Additionally, children in the H-SLI group performed much worse than their TD peers on the non-word repetition test. More specifically, the H-SLI group was less accurate at repeating three-syllable (F(1,29)=7.061, p=0.017) and four-syllable (F(1,29)=15.087, p=0.001) non-words, with a trend in the same direction for two-syllable non-words (F(1,29)=3.73, p=0.064). The Recalling Sentences sub-test of CELF and the non-word repetition test, which most clearly differentiated our participant groups, have been shown to have high sensitivity and specificity when identifying children with language disorders (Archibald & Gathercole, 2006; Conti-Ramsden, Botting, & Faragher, 2001; Dollaghan & Campbell, 1998; Ellis Weismer et al., 2000; Graf Estes, Evans, & Else-Quest, 2007; Hesketh & Conti-Ramsden, 2013). Furthermore, parents of 8 H-SLI children noted the presence of at least some language-related difficulties in their children during the time of testing. Four H-SLI children were still receiving speech therapy, with 3 more H-SLI children having stopped speech therapy approximately a year earlier. We take these results to indicate that, despite scoring within the normal range on CELF, our H-SLI children continued to lag behind their TD peers in linguistic competence. Indeed, earlier research shows that school-age children with a history of SLI often continue to exhibit albeit more subtle but clear linguistic difficulties (Conti-Ramsden et al., 2001; Hesketh & Conti-Ramsden, 2013; Rice, Hoffman, & Wexler, 2009). Such subtle differences in language competence may contribute to the development of learning disorders in other domains, most notably reading (Catts, Fey, Tomblin, & Zhang, 2002; Hesketh & Conti-Ramsden, 2013; Scarborough & Dobrich, 1990). Lastly, scores within the normal range on standardized tests in children with an established earlier diagnosis of SLI may not by themselves be sufficient to establish a true recovery because they may hide atypical cognitive strategies used by these children during testing (Karmiloff-Smith, 2009). Because audiovisual processing has a protracted developmental course, it is important to understand how children with documented language learning difficulties during early childhood compare to their TD peers during school-age years on the ability to detect audiovisual temporal asynchrony. Therefore, an older group of children was selected for this study.

Table 1.

CELF-4 scores of the H-SLI and TD children

C&FD RS FS WS WC CLS

Sub SS %ile SS %ile SS %ile SS %ile R-SS R-%ile E-SS E-%ile T-SS T-%ile SS %ile
H-SLI
1 10 50 9 37 14 91 9 37 102 55
2 13 84 8 25 11 63 13 84 108 70
3 11 63 6 9 8 25 10 50 93 32
4 8 25 8 25 6 9 11 63 90 25
5 14 91 9 37 12 75 12 75 111 77
6 15 95 13 84 10 50 14 91 118 88
7 8 25 5 5 7 16 7 16 81 10
8 11 63 8 25 13 84 6 9 97 42
9 11 63 12 75 13 84 13 84 16 98 15 95 117 87
10 12 75 7 16 11 63 10 50 7 16 8 29 97 42
11 12 75 9 37 15 95 14 91 11 63 13 84 114 82
12 7 16 2 0.4 6 9 8 25 3 5 6 9 72 3
13 9 37 7 16 12 75 8 25 7 16 7 16 93 32
14 13 84 6 9 10 50 11 63 10 50 10 50 98 45
15 10 50 8 25 9 37 13 84 13 84 13 84 99 47

Mean 10.9 7.8 10.5 10.3 10.7 9.6 10.1 99.3

TD
1 14 91 12 75 15 95 12 75 120 91
2 12 75 12 75 11 63 9 37 106 66
3 14 91 9 37 14 91 10 50 111 77
4 11 63 12 75 13 84 13 84 114 82
5 13 84 11 63 13 84 10 50 111 77
6 14 91 15 95 13 84 12 75 121 92
7 13 84 15 95 13 84 13 84 121 92
8 13 84 11 63 12 75 13 84 13 84 13 84 114 82
9 11 63 12 75 15 95 16 98 17 99 17 99 123 94
10 12 75 9 37 10 50 13 84 12 75 13 84 106 66
11 7 16 8 25 14 91 10 50 13 84 12 75 100 50
12 10 50 13 84 10 50 13 84 11 63 12 75 108 70
13 13 84 12 75 13 84 10 50 8 25 9 37 111 77
14 12 75 9 37 8 25 11 63 11 63 11 63 99 47
15 11 63 10 50 13 84 11 63 8 25 9 37 104 61

Mean 12.0 11.3 12.5 11.3 12.1 11.6 12.0 111.3

p .174 .000 .032 .407 .228 .295 .245 .005

C&FD = Concepts and Following Directions; RS = Recalling Sentences; FS = Formulated Sentences; WS = Word Structure; WC = Word Classes; R = Receptive; E = expressive; T = Total; CLS = Core Language Score; SS = standardized score; %ile = percentile. p values represent a group comparison based on a one-way ANOVA.

All participants were free of neurological disorders, passed a hearing screening at a level of 20 dB HL at 500, 1000, 2000, 3000, and 4000 Hz, reported to have normal or corrected-to-normal vision, and were not taking medications that may affect brain function (such as anti-depressants) at the time of study. All had normal non-verbal intelligence and showed no signs of autism spectrum disorders. One child in the H-SLI group had a diagnosis of ADHD, but was not taking medications at the time of study. Additionally, two more H-SLI children fell into the strongly atypical range based on the ADHD Index of the Parent Rating Scale of the Conners’ Rating Scales – Revised. The ADHD Index of the Conners’ Questionnaire is based on the set of questions that have been shown to be most effective at distinguishing children at risk for ADHD from their TD peers (Conners, 1997). Children with either a diagnosis of ADHD or a high ADHD index on the Conners’ Questionnaire were not excluded from the study because the diagnosis of SLI preceded that of ADHD and because recent studies suggest that linguistic problems associated with ADHD are qualitatively different from those observed in SLI (Redmond, 2005; Redmond, Thompson, & Goldstein, 2011). However, we included information on the status of attention in both groups of children, together with several other variables, as a predictor in a multiple regression analysis in order to understand its role in performance on the SJT. One child in the H-SLI group had a diagnosis of dyslexia, based on parental report. All children in the TD group had language skills within the normal range and were free of ADHD. None of the adult participants had a history of SLI or ADHD based on self-report.

Tables 2 and 3 provide group means and standard errors (SE) for the two groups of children for all aforementioned measures (except for results for CELF4 which are shown in Table 1). The H-SLI children did not differ from TD peers in age, non-verbal intelligence, or SES as measured by mother’s years of education. Although significant, the difference in father’s years of education between the two groups was less than three years. As mentioned earlier, children in the H-SLI group performed significantly worse on the Formulated Sentences and Recalling Sentences sub-tests of CELF and on the non-word repetition test. In addition, this group of children scored significantly lower on both the Number Memory Forward (group, F(1,29)=8.885, p=0.006) and the Number Memory Reversed (group, F(1,29)=6.546, p=0.016) sub-tests of TAPS-3. Based on the Laterality Index of the Edinburgh Handedness Questionnaire, one child in the H-SLI group was left-handed. All other participants were right-handed.

Table 2.

Group means for age, TONI-4, CARS-2, ADHD Index, SES, and TAPS-3

Age (years, months) TONI-4 CARS-2 ADHD Indexa Mother’s Education (years)b Father’s Education (years)b TAPS-3
Number Memory
Forward Reversed

H-SLI 9;0(0.3) 108.3(2.4) 15.4(0.1) 52.5(2.8) 16.6(0.8) 14.9(0.4) 7.8(0.4) 9.3(0.7)
TD 9;1(0.3) 107.6(2.1) 15.3(0.2) 48.0(1.6) 16.3(0.8) 17.5(0.8) 10.0(0.6) 11.7(0.6)
p 0.921 0.821 0.590 0.173 0.819 0.012 0.006 0.016

Note. Numbers for TONI-4, CARS-2, ADHD Index, and the TAPS-3 subtests represent standardized scores. Numbers in parentheses are standard errors of the mean. P-values reflect a group comparison. TONI–4 = Test of Nonverbal Intelligence—4th edition; CARS–2 = Childhood Autism Rating Scale—2nd edition; ADHD = attention deficit/hyperactivity disorder; TAPS–3 = Test of Auditory Processing Skills—3rd edition.

a

From the Parent Rating Scale of the Conners’ Rating Scales—Revised.

b

Measure of children’s SES.

Table 3.

Group means for nonword repetition and TEA-Ch

Nonword Repetition TEA-Ch

Syllables
Sky Search Sky Search Dual Task Map Mission
1 2 3 4 Total


H-SLI 91.1(1.9) 95.3(1.1) 88.1(3.2) 64.0(2.8) 80.9(1.7) 9.5(0.8) 7.4(0.7) 7.9(0.8)
TD 92.3(1.3) 98.3(1.1) 96.9(0.9) 79.6(2.9) 90.1(1.2) 9.3(0.4) 8.8(0.9) 9.1(0.7)
p 0.619 0.064 0.017 0.001 0.000 0.882 0.220 0.284

Note. Numbers for TEA-Ch represent standardized scores. Numbers for Nonword Repetition are percent of correctly repeated phonemes. Numbers in parentheses are standard errors of the mean. P-values reflect a group comparison.

2.2 Stimuli and Experimental Design

All participants performed the same SJT task, which was presented as a game. Prior to the start of the experiment, participants watched a video with instructions. They were told that they would see a dragon in the middle of the screen and pictures of a boy and a girl holding futuristic-looking weapons at the top left and right corners of the screen (see Figure 1). They were also told that these characters lived on the planet Cabula, where dragons often raid apple orchards and steal apples. Dragons can only be scared away by special weapons. One weapon shoots lights while another shoots sounds. In order for dragons to be scared away, the light and the sound need to occur at exactly the same time. If either the light or the sound is delayed or one of the children forgets to shoot (and hence there is only the light but no sound or there is only the sound but no light), dragons do not get scared away. Participants were then asked to indicate by a button press on a response box (RB-530, Cedrus Corporation) during each trial whether or not the dragon was scared away (i.e., whether the sound and the light occurred at the same time or not). The images of the girl and the boy with futuristic weapons were present during the instructions video and on the first screen of each block prior to the onset of test trials. To avoid a shift of attention away from the dragon, their images were not present while participants performed the task. Task instructions were kept identical for children and adults. The “light” stimulus was an explosion-shaped red figure that appeared over the dragon’s body. The “sound” stimulus was a 2 kHz pure tone. Both were 200 ms in duration. This combination of sound and light stimuli was presented at the following temporal offsets: 0, 100, 200, 300, 400, and 500 ms. In half the trials, the sound preceded the light (AV trials), while in another half, the light preceded the sound (VA trials). Additionally, there were visual only (VO) and auditory only (AO) trials, during which either only the explosion-shaped figure appeared without the accompanying tone or only the tone was presented without the accompanying explosion-shaped figure. Each experimental block contained 5 instances of each of the 13 types of trials (5 AV offsets, 5 VA offsets, synchronous presentation, AO, and VO trials), presented in a random order. Each participant completed 10 experimental blocks, which yielded 50 responses for each trial type. Presentation of trials was controlled by the Presentation software (www.neurobs.com). Participants were instructed to press one button for trials on which the Dragon was scared away (synchronous stimulus presentations) and another for trials on which the Dragon was not scared away (asynchronous and AO/VO trials). Responses were recorded within a 2200 ms response window time-locked to the appearance of the first stimulus in each pair. The response window was followed by an inter-trial interval varying randomly among 4 values: 350ms, 700ms, 1050ms, and 1400ms. Hand to response button mapping was counterbalanced across participants.

Figure 1. Stimuli.

Figure 1

During the instructions video and just prior to the onset of each experimental block, participants saw a dragon in the middle of the screen and a boy and a girl holding futuristic looking weapons in the top right and left corners. To avoid the shift of attention away from the dragon (over whom the visual stimuli were presented), only the dragon was present on the screen during all experimental trials. The visual stimulus was a red explosion-shaped figure that appeared over the dragon’s body. The auditory stimulus was a 2 kHz pure tone.

At the beginning of the session, participants were fitted with an EEG cap and seated in a dimly-lit sound-attenuating booth, approximately 4 feet from a computer monitor. Sounds were presented at 60 dB SPL via a sound bar located directly under the monitor. Participants were first shown the instructions video and then practiced the task until it was clear. In order to make sure that children understood the task, they were prompted with a set of questions to describe under which conditions they were to press each button. Each block lasted approximately 3 minutes. Children played several rounds of the board game of their choice after each block and had a longer break after half of all blocks were completed. Adult participants had short breaks after each block and a longer one in the middle. Together with game breaks, the testing session lasted approximately 1 hour.

2.3 ERP Recordings and Data Analysis

Electroencephalographic (EEG) data were recorded from the scalp at a sampling rate of 512 Hz using 32 active Ag-AgCl electrodes secured in an elastic cap (Electro-Cap International Inc., USA). Electrodes were positioned over homologous locations across the two hemispheres according to the criteria of the International 10-10 system (American Electroencephalographic Society, 1994). The specific locations were as follows: midline sites: FZ, CZ, PZ, and OZ; mid-lateral sites: FP1/FP2, AF3/AF4, F3/F4, FC1/FC2, C3/C4, CP1/CP2, P3/P4, PO3/PO4, and O1/O2; and lateral sites: F7/F8, FC5/FC6, T7/T8, CP5/CP6, and P7/P8; and left and right mastoids. EEG recordings were made with the Active-Two System (BioSemi Instrumentation, Netherlands), in which the Common Mode Sense (CMS) active electrode and the Driven Right Leg (DRL) passive electrode replace the traditional “ground” electrode (Metting van Rijn, Peper, & Grimbergen, 1990). During recording, data were displayed in relationship to the CMS electrode and then referenced offline to the average of the left and right mastoids (Luck, 2005). The Active-Two System allows EEG recording with high impedances by amplifying the signal directly at the electrode (BioSemi, 2013; Metting van Rijn, Kuiper, Dankers, & Grimbergen, 1996). In order to monitor for eye movement, additional electrodes were placed over the right and left outer canthi (horizontal eye movement) and below the left eye (vertical eye movement). Horizontal eye sensors were referenced to each other, while the sensor below the left eye was referenced to FP1 in order to create electro-oculograms. Prior to data analysis, EEG recordings were filtered between 0.1 and 30Hz. Individual EEG records were visually inspected to exclude trials containing excessive muscular and other non-ocular artifacts. Additionally, all audiovisually synchronous and VO trials in which blinks occurred during stimulus presentation were excluded. Ocular artifacts were corrected by applying a spatial filter (EMSE Data Editor, Source Signal Imaging Inc., USA) (Pflieger, 2001). Similar to the independent component analysis (ICA), this method is able to separate ocular artifacts from brain activity without constructing a head model and without subtracting ocular activity from EEG channels. However, unlike the ICA, it bypasses the need to subjectively determine which components need to be excluded by relying on a representative set of artifacts and clean data segments. ERPs were epoched starting at 200 ms pre-stimulus and ending at 1000 ms post-stimulus onset. On trials in which the auditory and visual stimuli were not completely synchronized, the stimulus onset was the onset of the first stimulus in a pair (for example, when the sound preceded the appearance of the explosion-shaped figure, ERPs were averaged to the onset of the sound). The 200 ms prior to the recording onset served as a baseline.

The goal of the ERP analyses was to examine how auditory and visual neural responses differed depending on the group membership and offset. However, because offsets were short, auditory and visual ERPs partially overlapped. In order to better isolate ERPs unique to each sensory modality, the following procedure was used. ERPs elicited by AO trials were subtracted from ERPs elicited by AV stimulus pairs (in which the sound preceded the light) and from those elicited by synchronous pairs, thus revealing ERPs specific to visual stimulation at each offset. Similarly, ERPs elicited by VO trials were subtracted from ERPs elicited by VA stimulus pairs (in which the light preceded the sound) and from ERPs elicited by synchronous pairs, revealing ERPs specific to auditory stimulation at each offset. All ERP analyses were conducted on the resulting difference waves. For the sake of clarity, all difference waves are labeled by referring to the specific subtraction that was performed to produce them. For example, the difference wave that resulted from AO ERPs being subtracted from the ERPs elicited by AV stimuli with a 100 ms offset is labeled AV100-A. AV0 and VA0 labels both refer to the completely synchronous audiovisual presentation. This method for isolating modality-specific ERPs is similar to that used previously by Sketelenburger and Vroomen (Stekelenburg & Vroomen, 2005).

Because the focus of the ERP analysis was on sensory processing of audiovisual stimuli, we measured only the P1 and N1 ERP components elicited by auditory and visual stimuli. For the ease of reference, auditory P1 and N1 are listed with an “a” preceding them (e.g., aP1 and aN1), while visual P1 and N1 are listed with a “v” preceding them (e.g., vP1 and vN1). Since difference waves have relatively lower signal-to-noise ratio compared to original waveforms, which can produce spurious peak latency measurements, we chose to analyze a 50 percent area latency measure and the corresponding amplitude measure for each of the above components (Luck, 2005). The 50 percent area latency measure refers to the point in the ERP waveform that divides the area under the waveform in half for a specified temporal window. The amplitude at such a point was used as a given component’s amplitude measure. Temporal windows for area peak and latency measurements were based on the visual inspection of AV0-A and VA0-V difference waves. Temporal windows for other offsets were calculated by adding 100–500 ms to the beginning and the end of each window. As an example, the visual P1 elicited in children was measured between 92 and 150 ms post-stimulus onset in the AV0-A condition, between 192 and 250 ms post-stimulus onset in the AV100-A condition, and between 292 and 350 ms post-stimulus onset in the AV200-A condition, etc., in order to properly capture the ERPs elicited by the incrementally delayed visual stimulus. In agreement with earlier reports, adults’ ERPs had noticeably shorter latencies. It was therefore impossible to keep the same temporal windows for measuring ERPs elicited in children and adults. Importantly, however, although temporal windows used for ERP measurements in adults were placed over an earlier portion of the waveform, their absolute width was kept identical for children and adults (see Table 4). Measurements from all 3 groups were included into the analyses of components’ amplitudes. However, due to differences in the temporal windows for measuring ERP components, latencies of P1 and N1 were compared only between the two groups of children.

Table 4.

Measured components: sites and temporal windows

Temporal Windows for Visual Components
Component Sites Sync-A AV100-A AV200-A AV300-A AV400-A AV500-A
Children
vP1 O1/2 92–150 192–250 292–350 392–450 492–550 592–650
vN1 O1/2 152–208 252–308 352–408 452–508 552–608 652–708

Adults
vP1 O1/2 66–124 166–224 266–324 366–424 466–524 566–624
vN1 O1/2 134–190 234–290 334–390 434–490 534–590 634–690
Temporal Windows for Auditory Components
Component Sites Sync-V VA100-V VA200-V VA300-V VA400-V VA500-V
Children
aP1 F3/4, FC1/2, C3/4, CP1/2, P3/4; FZ, CZ, PZ 50–98 150–198 250–298 350–398 450–498 550–598
aN1 F3/4, FC1/2, C3/4, CP1/2, P3/4; FZ, CZ, PZ 100–152 200–252 300–352 400–452 500–552 600–652

Adults
aP1 F3/4, FC1/2, C3/4, CP1/2, P3/4; FZ, CZ, PZ 26–74 126–174 226–274 326–374 426–474 526–574
aN1 F3/4, FC1/2, C3/4, CP1/2, P3/4; FZ, CZ, PZ 100–152 200–252 300–352 400–452 500–552 600–652

Note. Difference waves from which measurements were taken are labeled with the specific subtraction used to produce them. For example, AV100-A is a difference wave that resulted from subtracting ERPs elicited by auditory only trials from ERPs elicited by an audiovisual stimulus in which the onset of the sound preceded the appearance of the explosion-shaped figure by 100 ms. All temporal windows are shown in milliseconds post-stimulus onset. Site abbreviations: F = frontal, FC = fronto-central, CP = centro-parietal, P = parietal, C = central, O = occipital. Odd numbers indicate electrodes positioned over the left hemisphere of the scalp, while even numbers indicate electrodes positioned over the right hemisphere of the scalp. FZ, CZ, and PZ are all midline sites.

Repeated-measures ANOVAs were used to evaluate behavioral and ERP results. For behavioral measures, ANOVAs included the factors of group (H-SLI and TD) and offset (100, 200, 300, 400, and 500). The following factors were used for ERP analyses over mid-lateral sites: group (H-SLI and TD), offset (0, 100, 200, 300, 400, and 500), hemisphere (O1 vs. O2 for visual components; F3, FC1, C3, CP1, P3 vs. F4, FC2, C4, CP2, P4 for auditory components), and site (for auditory components only since in visual analyses each hemisphere was represented by one site; F3 and F4 vs. FC1 and FC2 vs. C3 and C4 vs. CP1 and CP2 vs. P3 and P4). Auditory ERPs were also evaluated over midline sites in a separate analysis with the following factors: group (H-SLI and TD), offset (0, 100, 200, 300, 400, and 500), and site (FZ vs. CZ vs. PZ). Because previous research has consistently reported a difference in individuals’ ability to detect AV and VA asynchronies (Bushara et al., 2001; Dixon & Spitz, 1980; Grant et al., 2004; Lewkowicz, 1996; McGrath & Summerfield, 1985; Van Wassenhove et al., 2007), separate ANOVAs were conducted on behavioral and ERP measures to AV and VA offsets. Significant main effects with more than two levels were evaluated with a Bonferroni post-hoc analysis. In cases where the omnibus analysis produced a significant interaction, it was further evaluated with step-down ANOVAs or t-tests, with factors specific to any given interaction. For all repeated measures with greater than one degree of freedom in the numerator, the Huynh-Feldt (H-F) adjusted p-values were used to determine significance (Hays, 1994). Effect sizes, indexed by the partial eta squared statistic (ηp2), are reported for all significant repeated-measures ANOVA effects. Because the goal of statistical analyses was to identify the influence of group membership on behavioral and neural response to audiovisual stimuli, we report significant results only for the factor of group and its interactions with other factors.

2.4 Regressions

In order to examine whether children’s audiovisual temporal discrimination abilities can predict language skills, we conducted separate linear regression analyses between performance on the SJT (as a predictor variable) and the following measures of linguistic aptitude (as response variables): the Concepts and Following Directions, Formulated Sentences, and Recalling Sentences subtests of CELF-4 as well as the CLS composite, and the 4-syllable nonword repetition. We performed a regression analysis not only on individual sub-tests of CELF-4 but also on the composite score because while the influence of audiovisual discrimination may not be strong enough to manifest itself in performance on individual tasks, it may nonetheless significantly predict a child’s overall language ability. Furthermore, the CLS composite incorporates performance on Word Classes-2 and Word Structure sub-tests, which could not be analyzed separately due to a small number of children to whom each test was administered. Group means differed significantly for several of the above language measures. As a result, we conducted regression analyses separately for the TD and H-SLI groups.

Additionally, because performance on the SJT has been shown to improve with age and required sustained attention, we conducted a multiple regression analysis with participants’ age, group (TD or H-SLI), the ADHD index, and the three sub-tests of TEA-Ch as the predictors and the SJT accuracy as the response variable in order to gauge the degree to which these variables contributed to children’s performance on the SJT task. A stepwise regression was performed to arrive at the final model. The SJT measure used for all regressions was the average of synchronous (i.e., erroneous) perceptions over the two longest offsets (400 and 500 ms). These offsets were selected because they showed the largest individual variability and the largest group differences. Regression analyses were conducted separately on AV and VA offsets.

3. Results

3.1. Behavioral Results

Behavioral results are shown in Figure 2. We evaluated children’s performance on the SJT by calculating the percent of trials at each offset for which stimuli were judged as synchronous, with an expectation that as the offset between visual and auditory modalities increased, the number of synchronous perceptions would decrease. Analysis of responses to VA and AV stimuli were nearly identical. Both yielded a significant effect of offset (VA: F(4,168)=212.809, p<0.001, ηp2=0.835; AV: F(4,168)=183.222, p<0.001, ηp2=0.814), with participants perceiving significantly fewer synchronous presentations as the offset increased from 100 to 200 ms (VA, p<0.001; AV, p<0.001), from 200 to 300 ms (VA, p<0.001; AV, p<0.001), from 300 to 400 ms (VA, p<0.001; AV, p<0.001), and from 400 to 500 ms (VA, p<0.001; AV, p=0.001), based on Bonferroni pairwise comparisons. The effect of group was also significant (VA, F(1,42)=26.621, p<0.001; AV, F(1,42)=30.91, p<0.001), with adults being more accurate than either group of children and with the TD children being more accurate than the H-SLI children (all p-values <0.01). The effects of offset and group were further defined by an offset by group interaction (VA: F(8, 168)=17.117, p<0.001, ηp2=0.449; AV: F(8, 168)=9.866, p<0.001, ηp2=0.32). Follow-up tests revealed a significant effect of group at 200, 300, 400, and 500 ms offsets for both AV and VA stimuli (F(2,42)=9.729–43.809, all p-values <0.001). Specific group differences based on Bonferroni pairwise comparisons are shown in Table 5 (top panel). Children in the H-SLI group were significantly less accurate than their TD peers when making judgments about 300, 400, and 500 ms offset in both AV and VA conditions (with the only exception being the VA300 stimulus, for which group difference fell just short of significance). The H-SLI children performed worse than adults on all but the shortest offsets (AV100 and VA100). Just as importantly, TD children, while more accurate than their H-SLI counterparts, nonetheless had significantly more difficulty at detecting audiovisual asynchrony than adults on all but three offsets (AV100, VA100, and VA500), revealing a prolonged developmental course for audiovisual temporal function.

Figure 2. Performance on the simultaneity judgment task.

Figure 2

The x axis displays all temporal offsets used in the study, with visual-auditory (VA) offsets in the left part of the graph and the audio-visual (AV) offsets in the right part of the graph. Numbers 1 through 5 correspond to temporal offsets of 100 ms to 500 ms, in 100 ms increments. For example, AV3 identifies a temporal offset in which the sound preceded the light by 300 ms. The y axis shows percent of synchronous perceptions. As offsets increased, percent of synchronous perceptions was expected to decrease. Error bars are standard errors of the mean.

Table 5.

Group differences on the simultaneity judgment task

n=15
VA500 VA400 VA300 VA200 VA100 SYNC AV100 AV200 AV300 AV400 AV500
H-SLI<TD *** *** NS NS NS NS NS NS ** *** ***
TD<Adults NS *** *** * NS NS NS ** *** ** *
H-SLI<Adults *** *** *** *** NS NS NS *** *** *** ***
n=12
VA500 VA400 VA300 VA200 VA100 SYNC AV100 AV200 AV300 AV400 AV500
H-SLI<TD ** * NS NS NS NS NS NS NS * *
TD<Adults NS *** *** ** NS NS NS ** *** ** *
H-SLI<Adults *** *** *** ** NS NS NS *** *** *** ***
*

p<0.05,

**

p<0.01,

***

p<0.001

NS=not significant

The top panel lists group differences for the analysis that included all 15 participants in each group. The bottom panel lists group differences for the analyses that were conducted only on the H-SLI children without any known attention difficulties and corresponding TD children and adults.

Because three of the H-SLI children either had a diagnosis of ADHD or scored high on the Conners’ ADHD index, we wanted to make sure that these children’s performance on the SJT did not skew our group results. We therefore re-ran our analyses with 12 H-SLI children without known attention difficulties, 12 age-matched TD children, and 12 adults. Similarly to the analysis with 15 participants in each group, a group by offset interaction was significant (VA, F(8,132)=13.61, p<0.001, ηp2=0.452; AV, F(8,132)=7.51, p<0.001, ηp2=0.313). Follow up analyses showed that group differences still remained, albeit over a somewhat smaller range of offsets (see Table 5, bottom panel).

Groups did not differ in their responses to completely synchronous presentations. However, analyses of accuracy to AO and VO stimuli yielded a significant effect of group (AO, F(2,44)=7.291, p=0.002; VO, F(2,44)=12.197, p<0.001), with H-SLI children being less accurate than both TD children and adults, based on Bonferroni pairwise comparisons (AO, H-SLI vs. TD, p=0.051, H-SLI vs. adults, p=0.002; VO, H-SLI vs. TD, p=0.004, H-SLI vs. adults, p<0.001). TD children and adults did not differ from each other. Additionally, in order to make sure that the offset and group effects were due to errors in stimulus identification rather than due to the lack of responses, we also analyzed the number of misses in each group. The VA analysis yielded a significant group by offset interaction (F(8,168)=2.473, p=0.022, ηp2=0.105), with TD children making more misses than adults to the 200, 300, 400, and 500 ms offsets (in all cases, p<0.05). The TD group also had more misses than the H-SLI group to the 400 ms offset (p=0.01). The AV analysis showed the effect of group (F(2,42)=8.389, p=0.001, ηp2=0.285), with adults having fewer misses than either TD children (p=0.001) or H-SLI children (p=0.015). No difference between the two groups of children was found. Although groups differed significantly in some comparisons, such differences were on average between only 1 and 7 misses.

Visual inspection of behavioral SJT data revealed that there was much larger individual variability in the H-SLI group, with some children falling within the TD range on the SJT and some consistently unable to detect asynchrony even at the longest offsets. Individual variability in each group is shown in Figure 3.

Figure 3. Individual differences in performance on the simultaneity judgment task.

Figure 3

The x axis displays all temporal offsets used in the study, with VA offsets in the left part of the graph and the AV offsets in the right part of the graph. Numbers 1 through 5 correspond to temporal offsets of 100 to 500 ms, in 100 ms increments. For example, AV3 identifies a temporal offset in which the sound preceded the light by 300 ms. The y axis shows percent of synchronous perceptions. Each line represents a separate participant.

3.1.1. Regressions

Figure 4 shows all significant regressions between the SJT and language measures, separately for the AV and VA conditions. In the H-SLI group, better performance on the SJT was associated with better scores on the C&FD subtest of CELF and with an overall higher CLS index, with a trend in the same direction for the FS subtest. This relationship held for both the AV and the VA offsets. In the TD group, better performance on the SJT appeared to be associated with worse scores on the FS subtest and the CLS index. However, further examination of the data strongly suggested that this paradoxical relationship was due to the fact that the younger children in the TD group had a higher CLS than the older children (e.g., there was a significant negative correlation between the CLS and age in the TD group: R=.561, F(1,14)=5.976, p=0.03), but had more difficulty with the SJT than their older counterparts. In a separate study in our laboratory, we tested seventeen 7–8 year-old and nineteen 10–11 year-old TD children on the same SJT (Kaganovich, in preparation). In neither group of children was worse performance on the SJT associated with better CELF-4 scores. These results suggest a relationship between audiovisual temporal function and language aptitude in H-SLI children. However, given a small number of participants in each group, a replication of these findings is needed.

Figure 4. Regressions between performance on the simultaneity judgment task and language tests.

Figure 4

The language measures that showed a significant relationship with the SJT were the Concepts and Following Directions (top panel) and the Formulated Sentences (middle panel) subtests of CELF-4, as well as the composite language score of CELF-4 (bottom panel). The y axis is the average percent of synchronous perceptions over the two longest offsets of 400 and 500 ms.

The results of the multiple regression analysis that aimed to predict performance on the SJT using group membership, age, the ADHD index, and the TEA-Ch scores are summarized in Table 6. Only group, age, and the ADHD index had significant enough p-values (<.05) to be included into the model. The final model was able to account for approximately 80% of the variability (R2 =0.799) in the performance on the AV offsets of the SJT and for approximately 60% of the variability (R2 =0.593) in the performance on the VA offsets of the SJT. The order in which variables were selected for the model was different for the AV and VA analyses, with the ADHD index entering before the group membership for performance on the AV offsets and with the group membership entering before the ADHD index for the VA offsets. More importantly, the regression coefficients for the predictors were similar for the AV and VA analyses. The ADHD index had a positive coefficient, and age had a negative coefficient. Thus, higher ADHD index scores were associated with worse performance on the SJT while older age was associated with better performance on the SJT. Importantly, in both analyses being a member of the H-SLI group was associated with worse performance on the SJT.

Table 6.

Multiple regression results

AV offset
Model Variable B SE Beta p R2 R2 change
1 Constant −50.439 19.864 .017 .416 .416
ADHD index 1.741 .645 .000
2 Constant 3.666 19.617 .853 .663 .246
ADHD index 1.386 .312 .514 .000
Group −24.207 5.45 −.514 .000
3 Constant 65.247 21.304 .005 .799 .136
ADHD index 1.519 .248 .563 .000
Group −23.282 4.295 −.494 .000
Age −7.662 1.827 −.372 .000
VA offset
Model Variable B SE Beta p R2 R2 change
1 Constant 70.467 10.684 .000 .33 .33
Group −25.067 6.757 −.574 .001
2 Constant 20.545 23.319 .386 .445 .115
Group −21.15 6.479 −.484 .003
ADHD index .877 .371 .351 .025
3 Constant 80.023 28.085 .008 .593 .148
Group −20.256 5.662 −.464 .001
ADHD index 1.005 .326 .402 .005
Age −7.4 2.408 −.388 .005

B = regression coefficient, SE = Standard Error

3.2 ERP Results

ERPs elicited by AO, VO, and audiovisually synchronous stimuli are shown in Figure 5. ERP waveforms elicited by these stimuli in children and adults differed in peaks’ latency, amplitude, and overall morphology. Generally speaking, adults’ components peaked earlier than children’s. Adults’ P1 was generally smaller while adults’ N1 larger in amplitude than the corresponding components in children’s waveforms. Lastly, children’s responses to visual and audiovisual stimuli contained an anterior negativity at approximately 400 ms post-stimulus onset, which was absent in adults’ waveforms.

Figure 5. ERPs elicited by AO, VO, and synchronous stimuli.

Figure 5

ERPs elicited by AO, VO, and synchronous audiovisual stimuli are shown for the H-SLI, TD, and adult groups over two representative electrodes. Negative is plotted up.

ERPs elicited by the visual part of audiovisual stimulation are shown in Figure 6, with measured components marked on the O2 site. Analysis of vP1 showed a larger amplitude of this component in the H-SLI children compared to adults at 200, 300, and 400 ms offsets over the right hemisphere (group by offset by hemisphere, (F(10,210)=1.944, p=0.041, ηp2=0.085; H-SLI vs. adults, all p<0.05 based on Bonferroni pairwise comparison). A similar finding for the 500 ms offset fell just short of significance (H-SLI vs. adults, p=0.053). The vP1 amplitude was larger in H-SLI than in TD children only at the 300 ms offset (p=0.034). Additionally, in the H-SLI group vP1 amplitude was significantly larger over the right compared to the left occipital site for 0, 300, 400, and 500 ms offsets (in all cases, p<0.05), with a trend in the same direction for the 200 ms offset (p=0.085). This hemispheric effect was absent in TD children and adults (see Figure 7). Analysis of the vN1 amplitude indicated that it was significantly smaller over the right than over the left occipital site only in the H-SLI group (group by hemisphere, F(2,42)=9.498, p<0.001, ηp2=0.311; hemisphere in H-SLI, F(1,14)=15.218, p=0.002, ηp2=0.521) This effect did not interact with offset. Analyses of vP1 and vN1 latencies did not yield any significant group differences between TD and H-SLI children.

Figure 6. Visual ERPs.

Figure 6

Visual ERPs elicited by the appearance of the explosion-shaped figure are shown for the H-SLI, TD, and adult groups over a representative set of electrodes. All ERPs are difference waves that were created by subtracting auditory only trials from AV stimuli with different offsets in order to reveal brain activity specific to visual stimulation. Difference waves are labeled based on the subtraction that was used to create them. For example, the AV500-A panel shows visual ERPs that were revealed by subtracting auditory only trials from ERPs elicited by audiovisual stimuli with a 500 ms offset. Measured components are marked on the O2 site. Negative is plotted up.

Figure 7. Hemispheric asymmetry of visual ERPs.

Figure 7

Visual ERPs elicited over the left hemisphere (LH) and right hemisphere (RH) occipital sites are overlaid for the three groups of participants and for each of the six offsets used in the study. The sustained positivity over the RH occipital site in the H-SLI group started at the onset of the vP1 component and continued for approximately 500 ms. Other visual components are not marked.

ERPs elicited by the auditory part of audiovisual stimulation are shown in Figure 8, with measured components marked on the C3 site. Several earlier reports showed that an adult-like aN1 may not appear in children’s ERPs until late childhood when sounds are presented at short inter-stimulus interval (e.g., Sharma et al., 1997; Sussman et al., 2008). However, because stimuli presented at short offsets in our study were from different modalities, both groups of children exhibited clear aP1 and aN1 components to sound onsets. Analysis of aP1 amplitude yielded no significant results. The aP1 latency, however, was overall shorter in the H-SLI group compared TD children (mid-lateral sites: group, F(1,28)=21.278, p<0.001, ηp2=0.432; midline sites: group, F(1,28)=14.626, p=0.001, ηp2=0.343). This effect interacted with offset (mid-lateral sites: group by offset, F(5,140)=34.947, p<0.001, ηp2=0.555; midline sites, F(5,140)=19.546, p<0.001, ηp2=0.411), with shorter latency in the H-SLI group for the 400 ms offset (mid-lateral sites: F(1,28)=199.758, p<0.001, ηp2=0.877; midline sites, F(1,28)=176.532, p<0.001, ηp2=0.863) and a trend in the same direction for the 500 ms offset (mid-lateral sites, (F(1,28)=2.757, p=0.108, ηp2=0.09; midline sites, F(1,28)=3.864, p=0.059, ηp2=0.121). Analysis of aN1 amplitude revealed that it was significantly enlarged in adults compared to H-SLI children and marginally enlarged compared to TD children at 100 ms offset (mid-lateral: group by offset, F(10,210)=2.359, p=0.014, ηp2=0.101, group, F(2,42)=6.504, ηp2=0.236; H-SLI vs. adults, p=0.003, TD vs. adults, p=0.064; midline: group by offset, F(10,210)=2.202, p=0.023, ηp2=0.095; group, F(2,42)=7.911, p=0.001, ηp2=0.274; H-SLI vs. adults, p=0.001, TD vs. adults, p=0.058) and that it was larger in TD children than in H-SLI children at the 500 ms offset (mid-lateral: group, F(2,42)=3.242, p=0.049; H-SLI vs. TD, p=0.047; midline: group, F(2,42)=2.843, p=0.069, ηp2=0.119; H-SLI vs. TD, p=0.068). Finally, analysis of the aN1 latency showed that the H-SLI group had marginally shorter latencies compared to their TD peers over the midline sites only (F(1,28)=3.910, p=0.058, ηp2=0.123). Insert Figure 8 here

Figure 8. Auditory ERPs.

Figure 8

Auditory ERPs elicited by the 2 kHz pure tone are shown for the H-SLI, TD, and adult groups over a representative set of electrodes. All ERPs are difference waves that were created by subtracting visual only trials from VA stimuli with different offsets in order to reveal brain activity specific to auditory stimulation. Difference waves are labeled based on the subtraction that was used to create them. For example, the VA500-V panel shows auditory ERPs that were revealed by subtracting visual only trials from ERPs elicited by visual-auditory stimuli with a 500 ms offset. Measured components are marked on the C3 site. Negative is plotted up.

4. Discussion

This study examined sensitivity to audiovisual temporal asynchrony in TD and H-SLI school-age children and adults. We found that children in the H-SLI group were profoundly less sensitive to temporal separations between auditory and visual modalities than their TD peers and adults, both when the sound preceded the visual stimulus and when the visual stimulus preceded the sound. These group differences persisted even when separations between modalities reached 400 and 500 ms. Just as importantly, TD children were less sensitive to audiovisual asynchrony than young adults, revealing a protracted developmental course of audiovisual temporal function. By using a multiple regression analysis we also showed that for children such factors as group membership (H-SLI or TD), age, and Conners’ ADHD index were able to account for 60% (the VA stimuli) to 80% (the AV stimuli) of variability in the SJT performance, with the H-SLI status, higher ADHD index (even in the absence of an official ADHD diagnosis), and younger age being predictive of poor audiovisual temporal function.

A separate analysis of a relationship between the SJT performance and language aptitude revealed that those H-SLI children who did better on the SJT also had higher scores on the C&FD subtest of CELF-4 and the overall language aptitude composite score, thus suggesting a relationship between audiovisual temporal discrimination and language proficiency in this group. This relationship was absent in the TD children. Given a relatively small number of participants in each group, this result requires replication. The importance of audiovisual processing for early phonological and lexical development has been well established. However, one might hypothesize that once these linguistic systems reach a certain level of maturity, visual information, and audiovisual temporal discrimination in particular, may play a smaller role. While language skills are relatively mature by mid-childhood in TD children, they often remain vulnerable well into adolescence and even adulthood in individuals with a history of SLI (Hesketh & Conti-Ramsden, 2013). Therefore, in the H-SLI group, a greater ability to use visual cues is likely to translate into better linguistic outcomes even during mid-childhood - the age range of our participants. Such prolonged dependence on visual information in children with SLI has also been reported by Botting and colleagues (Botting, Riches, Gaynor, & Morgan, 2010), who showed that children with SLI rely on gesture for speech comprehension significantly longer than their TD peers, likely due to weaker auditory speech comprehension skills in this group.

Alternatively, a correlation between the SJT and language aptitude in the H-SLI group may be due to some other cognitive function that plays a prominent role in both tasks. Previous research has demonstrated that children with SLI have difficulty with engaging, sustaining, and selectively focusing their attention (Dispaldro et al., in press; Finneran, Francis, & Leonard, 2009; Spaulding, Plante, & Vance, 2008; Stevens, Sanders, & Neville, 2006); therefore, a difference in attentional skills might have contributed to differences in group performance on both the SJT and CELF-4. To examine this possibility, we evaluated attention skills in both groups via two separate measures. The ADHD Index of the Conners’ Rating Scales – Revised was used to identify children who may be at risk for developing ADHD, while the TEA-Ch test directly examined children’s ability to sustain and divide attention. The two groups of children did not differ in either the ADHD index or the TEA-CH scores; however, based on the multiple regression analysis, the ADHD index accounted for a significant amount of variability in performance on the SJT. Nevertheless, after removing the two H-SLI children with the ADHD index in the highly atypical range and one H-SLI child with an official diagnosis of ADHD, from the SJD analysis, the group difference was still significant in both the AV and VA conditions, albeit over a somewhat smaller range of audiovisual offsets. In sum, attention does appear to have played some role in children’s performance on the SJT; however, group differences in the audiovisual temporal function cannot be fully accounted for by differences in attention skills.

Our results extend previous research on the status of rapid stimulus processing in SLI in an important way. Earlier reports suggest that individuals with SLI have difficulty with unimodal stimuli presented in close succession, but perform similarly to their TD peers when the rate of presentation slows down, provided that stimulus duration is sufficient (Tallal & Piercy, 1973a, 1973b). However, perception of multimodal temporal offsets appears to be qualitatively different, and the H-SLI group remained impaired even at the longest offsets. As mentioned in the Introduction, growing research shows that audiovisual temporal asynchrony is typically not perceived even by adults until the offset reaches tens or even hundreds of milliseconds (depending on the nature of the stimuli and other conditions), likely due to the integration of multisensory stimuli into a holistic percept. The amount of the temporal offset necessary for asynchrony perception is very large in infancy but becomes smaller with development. One interpretation of our findings is that the H-SLI children integrate audiovisual information over much longer temporal offsets compared to their TD peers. If true, their audiovisual integration skills may resemble those of significantly younger children. However, another interpretation is also possible, namely, that children with a history of SLI are generally poor at making temporal judgments. Because temporal intervals flanked by multimodal stimuli (such as the ones used in our study) are more difficult to process compared to temporal intervals flanked by unimodal stimuli (e.g., Grondin & Rousseau, 1991), the H-SLI children’s performance on the SJT remained poor even at relatively large offsets. Teasing these two possibilities apart is an important theoretical question that requires future studies.

The use of simple non-speech stimuli allowed us to examine audiovisual temporal discrimination without the potentially confounding factors of linguistic ability and sound complexity. However, the use of such stimuli raises the question of relevance of our results to audiovisual speech perception in SLI. Undoubtedly, future studies on audiovisual temporal processing of speech in SLI will provide a more complete picture of the status of audiovisual integration in this disorder. However, at least some reports suggest that the temporal processing of speech and non-speech may be interconnected. For example, deficits in rapid auditory processing in SLI have been shown for both speech and non-speech stimuli. Furthermore, in the study by Leonard and colleagues (Leonard, Ellis Weismer, et al., 2007), non-linguistic speed of processing has been found to be one of two factors that significantly predicted linguistic skills in a group of school-age children with language impairment and their TD peers. Deficits in rapid processing appear before the onset of spoken language and may potentially be a precursor to linguistic difficulties during language acquisition. For example, infants’ rapid auditory processing of non-linguistic stimuli has been shown to predict receptive and expressive language during the second and third year of life (Choudhury & Benasich, 2011).

As summarized in the Introduction, several previous studies reported a reduced susceptibility to the McGurk illusion in children and adults with SLI, which was thought to indicate atypical audiovisual processing in this population (Boliek et al., 2010; Norrix et al., 2006; Norrix et al., 2007). Because successful audiovisual integration depends on a close temporal alignment between different modalities, our results suggest that poor audiovisual temporal function in children with SLI may be a contributing factor to this group’s performance on the McGurk task. This proposition is supported by a recent study of Stevenson and colleagues who reported a significant correlation between the ability to detect audiovisual asynchrony in simple non-speech stimuli and susceptibility to the McGurk illusion in adults (Stevenson, Zemtsov, & Wallace, 2012), with individuals who could detect audiovisual asynchrony over smaller visual-auditory offsets being significantly more likely to report the perception of the fused McGurk syllables. However, the degree to which this relationship holds in children is not yet known.

In sum, future studies on audiovisual temporal processing of speech in SLI are needed. Because reliable detection of temporal audiovisual asynchrony in speech requires longer separations compared to non-speech stimuli (e.g., Dixon & Spitz, 1980), we would predict that children with SLI will show even greater deficits in audiovisual temporal function when asked to judge the audiovisual simultaneity of speech.

Our ERP analyses aimed to determine whether early neural responses to auditory and visual stimuli differed among groups as a function of a temporal offset between auditory and visual stimuli. All groups showed clear P1 and N1 components in auditory and visual modality, regardless of the length of the offset. Analysis of group differences yielded a complex picture, with some differences tied to specific offsets and others manifesting themselves across the entire range of stimuli. Because the majority of earlier studies focused on differences in auditory ERPs between TD and SLI children, we were surprised to find that by far the largest group differences in early sensory encoding were present in the visual modality. More specifically, early visual ERPs in the H-SLI children were characterized by sustained positivity which started during the time window of the visual P1 component and continued for approximately 500 ms post-stimulus onset, leading to a significant enhancement of vP1 and a reduction of vN1 in this group. Importantly, this positivity was observed only over the right hemisphere. During the vP1 window it was present in ERPs elicited by stimuli with 300, 400, and 500 ms offset as well as by the visual component of the synchronous stimulus, but was present in ERPs to all offsets during the vN1 temporal window (see Figure 7).

An enhanced right-hemisphere response to visual stimuli in H-SLI children may indicate an atypical hemispheric specialization for sensory visual processing. While atypical lateralization of language function has long been suggested as one of the markers of developmental language disorders (e.g., Bishop, 2013; de Guibert et al., 2011; Whitehouse & Bishop, 2008), much less is known about hemispheric dominance of non-linguistic abilities in this group of children, and future studies on this topic are greatly needed. Alternatively, an increased right hemisphere activation to flash stimuli in the H-SLI children may reflect a delayed, rather than frankly abnormal, development of sensory processing. Figure 9 shows the average of vP1 amplitude over all temporal offsets for the left (O1) and right (O2) occipital sites in the TD and H-SLI children and in adults. As can be seen, a substantial number of TD children had greater absolute values over the right occipital site; however, this hemispheric difference was consistently smaller than in the SLI group and almost completely absent in adults. Based on these data, one might hypothesize that the greater engagement of the right hemisphere by simple visual stimuli may be typical for younger TD children, but this asymmetry reduces with development. Both structural and functional hemispheric asymmetry is certainly pervasive in the developing brain (e.g., Dennis & Thompson, in press); however, its significance in many cases remains unknown. Studies of visual ERP responses across a broad range of ages during development are needed to provide further insight into this issue.

Figure 9. Hemispheric asymmetry of visual ERPs - individual differences.

Figure 9

The vP1 amplitude over the left hemisphere (O1) and right hemisphere (O2) occipital sites is shown for each participant. Data points represent the average of vP1 amplitude over the six temporal offsets.

It is noteworthy, that the H-SLI children also had significantly lower accuracy compared to TD children in response to VO stimuli. In fact, they responded correctly on just 67.5 percent of trials compared to 90.3 percent of trials in TD children. The two groups of children did not differ in the number of misses to VO stimuli, so the low accuracy in the H-SLI group could not be explained by simple inattention or slower reaction time. An interesting empirical question to be addressed by future studies is whether atypical neural response to visual stimulation in children with language learning difficulties is related to their ability to reliably detect and identify visual stimuli.

Analysis of auditory ERPs showed that the TD children had a larger aN1 compared to the H-SLI group at the 500 ms offset. Additionally, the latency of aP1 and aN1 was longer in the TD compared to the H-SLI children. In the case of aP1, this group difference was restricted to the two longest offsets but was present across-the-board over the midline sites during the aN1 time window. An examination of aN1 amplitudes elicited by sounds at different offsets in the TD and H-SLI groups revealed that TD children had a larger aN1 to the 500 ms (−9.521 μV) as compared to the 400 ms (−4.786 μV) offset while amplitudes in the H-SLI group did not change as a function of offset (−4.939 μV at 400 ms and −4.608 μV at 500 ms). The functional significance of this group difference remains to be determined. One possibility is that at the 500 ms offset, TD children are able to complete the processing of the preceding visual stimulus well before the onset of the auditory one, thus potentially allocating a greater amount of attention to the auditory signal, which could result in amplitude enhancement. The H-SLI group, on the other hand, may take longer to process the preceding visual stimulus even at the 500 ms offset rendering the following auditory tone less salient. This hypothesis would agree with low accuracy to VO trials and atypical ERPs to visual stimuli in the H-SLI group.

Despite several significant group differences described above, both auditory and visual stimuli elicited largely similar ERPs in TD and H-SLI children. Clear P1 and N1 components were observed in both modalities at all temporal offsets indicating that a neural response was sufficiently salient even at shortest separations. Several reported group effects were limited to longer temporal offsets suggesting that groups differed in at least some aspects of sensory encoding during audiovisual processing. However, such differences did not clearly parallel behavioral findings and thus were probably not a key factor in the H-SLI’s group reduced sensitivity to audiovisual asynchrony. Consequently, the main cause of the sluggish audiovisual temporal function in the H-SLI children is likely to reside in relatively late perceptual and cognitive processes that lie beyond early sensory encoding.

Interestingly, a difficulty in detecting audiovisual temporal asynchrony has also been found in autism and dyslexia (Foss-Feig et al., 2010; Hairston, Burdette, Flowers, Wood, & Wallace, 2005; Kwakye, Foss-Feig, Cascio, Stone, & Wallace, 2011; Stevenson et al., 2014). The relatively high co-morbidity of the three disorders (Conti-Ramsden, Simkin, & Botting, 2006; Flax et al., 2003) might suggest that impairment in the same neural mechanism contributes to the disruption of audiovisual temporal discrimination across the three populations. However, this interpretation should be viewed with caution. In a positron emission tomography study, Bushara and colleagues (Bushara et al., 2001) has shown that the task of detecting temporal asynchrony between auditory and visual stimuli engages a large neural network that includes insular, posterior parietal, prefrontal, and cerebellar areas of the brain, as well as the right insula. Therefore, hypothetically, a disruption in different parts of this network might result in the impairment of audiovisual temporal function.

Lastly, the results of this study show that children who do not currently hold a diagnosis of specific language impairment (i.e., whose language skills fall within the normal range on standardized tests), but who have a history of this disorder, may continue to differ significantly from their TD peers in a number of linguistic and non-linguistic cognitive functions, such as general linguistic ability, verbal working memory, and audiovisual temporal discrimination. Because in the field of speech-language pathology decisions about the usefulness of language therapy are often based on the outcome of standardized tests, our findings raise a concern that some children may stop receiving therapy before their difficulties truly resolve.

In sum, our findings show that, as a group, children with a history of SLI have significantly more difficulty at identifying audiovisual temporal asynchrony compared to their age-matched TD peers and adults, even when the two modalities are separated by as much as 500 ms. Importantly, those H-SLI children who are more sensitive to audiovisual asynchrony also have higher composite language scores, indicating a relationship between audiovisual temporal function and language aptitude in this group. Although TD children overall performed significantly better than children in the H-SLI group, they nonetheless were still less accurate than adults. Audiovisual temporal function thus appears to have a remarkably prolonged developmental course and is not yet mature by mid-childhood. Analyses of sensory ERPs elicited by audiovisual stimuli at different offsets revealed a number of group differences, some of which were mediated by the stimuli offset and some were general in nature, with the most pronounced group differences observed in visual ERPs. Detecting temporal correspondences between auditory and visual stimuli is one of the pre-requisites for successful audiovisual integration. Accumulating evidence suggests that audiovisual processing plays an important facilitative role in typical language acquisition and speech perception. Our findings therefore highlight the need for further evaluation of the role of atypical audiovisual integration in the development of SLI.

Acknowledgments

This research was supported in part by grants P30DC010745 and R03DC013151 from the National Institute on Deafness and Other Communicative Disorders, National Institutes of Health. The content is solely the responsibility of the authors and does not necessarily represent the official view of the National Institute on Deafness and Other Communicative Disorders or the National Institutes of Health. We are thankful to Dr. Patricia Deevy and Kevin Barlow for their assistance with different stages of this project and to children and their families for participation. We are also grateful to Caryn Herring for creating the picture of the dragon and to Mohneet Ahuja for assistance with statistical analyses.

References

  1. American Electroencephalographic Society. Guideline thirteen: Guidelines for standard electrode placement nomenclature. Journal of Clinical Neurophysiology. 1994;11:111–113. [PubMed] [Google Scholar]
  2. Archibald LMD, Gathercole SE. Nonword repetiton: A comparison of tests. Journal of Speech, Language, and Hearing Research. 2006;49:970–983. doi: 10.1044/1092-4388(2006/070). [DOI] [PubMed] [Google Scholar]
  3. Barutchu A, Danaher J, Crewther SG, Innes-Brown H, Shivdasani MN, Paolini AG. Audiovisual integration in noise by children and adults. Journal of Experimental Child Psychology. 2010;105:38–50. doi: 10.1016/j.jecp.2009.08.005. [DOI] [PubMed] [Google Scholar]
  4. Bergeson TR, Pisoni DB, Davis RAO. Development of audiovisual comprehension skills in prelingually deaf children with cochlear implants. Ear and Hearing. 2005;26(2):149–164. doi: 10.1097/00003446-200504000-00004. [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. BioSemi. Active Electrodes 2013. 2013 from http://www.biosemi.com/active_electrode.htm.
  6. Bishop DVM. Cerebral asymmetry and language development: Cause, correlate, or consequence? Science. 2013;(340) doi: 10.1126/science.1230531. [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Bishop DVM, McArthur GM. Immature cortical responses to auditory stimuli in specific language impairment: evidence from ERPs to rapid tone sequences. Developmental Science. 2004;7(4):F11–F18. doi: 10.1111/j.1467-7687.2004.00356.x. [DOI] [PubMed] [Google Scholar]
  8. Boliek CA, Keintz C, Norrix LW, Obrzut J. Auditory-visual perception of speech in children with leaning disabilities: The McGurk effect. Canadian Journal of Speech-Language Pathology and Audiology. 2010;34(2):124–131. [Google Scholar]
  9. Botting N, Riches N, Gaynor M, Morgan G. Gesture production and comprehension in children with specific language impairment. British Journal of Developmental Psychology. 2010;28(Pt 1):51–69. doi: 10.1348/026151009x482642. [DOI] [PubMed] [Google Scholar]
  10. Brandwein AB, Foxe JJ, Russo NN, Altschuler TS, Gomes H, Molholm S. The development of audiovisual multisensory integration across childhood and early adolescence: A high-density electrical mapping study. Cerebral Cortex. 2011;21(5):1042–1055. doi: 10.1093/cercor/bhq170. [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Bristow D, Dehaene-Lambertz G, Mattout J, Soares G, Glida T, Baillet S, Mangin JF. Hearing faces: How the infant brain matches the face it sees with the speech it hears. Journal of Cognitive Neuroscience. 2008;21(5):905–921. doi: 10.1162/jocn.2009.21076. [DOI] [PubMed] [Google Scholar]
  12. Brown L, Sherbenou RJ, Johnsen SK. Test of Nonverbal Intelligence. 4. Austin, Texas: Pro-Ed: An International Pubilsher; 2010. [Google Scholar]
  13. Burnham D, Dodd B. Auditory-visual speech integration by prelinguistic infants: Perception of an emergent consonant in the McGurk effect. Developmental Psychobiology. 2004;45(4):204–220. doi: 10.1002/dev.20032. [DOI] [PubMed] [Google Scholar]
  14. Bushara KO, Grafman J, Hallett M. Neural correlates of audio-visual stimulus onset asynchrony detection. The Journal of Neuroscience. 2001;21(1):300–304. doi: 10.1523/JNEUROSCI.21-01-00300.2001. [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Catts H. The relationship between speech-language impairments and reading disabilities. Journal of Speech, Language, and Hearing Research. 1993;36:948–958. doi: 10.1044/jshr.3605.948. [DOI] [PubMed] [Google Scholar]
  16. Catts H, Fey ME, Tomblin JB, Zhang X. A longitudinal investigation of reading outcomes in children with language impairments. Journal of Speech, Language, and Hearing Research. 2002;45:1142–1157. doi: 10.1044/1092-4388(2002/093). [DOI] [PubMed] [Google Scholar]
  17. Čeponienė R, Rinne T, Näätänen R. Maturation of cortical sound processing as indexed by event-related potentials. Clinical Neurophysiology. 2002;113:870–882. doi: 10.1016/s1388-2457(02)00078-0. [DOI] [PubMed] [Google Scholar]
  18. Choudhury N, Benasich AA. Maturation of auditory evoked potentials from 6 to 48 months: Prediction to 3 and 4 year language and cognitive abilities. Clinical Neurophysiology. 2011;122:320–338. doi: 10.1016/j.clinph.2010.05.035. [DOI] [PubMed] [Google Scholar]
  19. Cohen MS. Handedness Questionnaire. 2008 Retrieved 05/27/2013, 2013, from http://www.brainmapping.org/shared/Edinburgh.php#.
  20. Conners KC. Conners’ Rating Scales - Revised. North Tonawanda, NY: MHS; 1997. [Google Scholar]
  21. Conti-Ramsden G, Botting N, Faragher B. Psycholinguistic markers for Specific Language Impairment (SLI) Journal of Child Psychology and Psychiatry. 2001;42(6):741–748. doi: 10.1111/1469-7610.00770. [DOI] [PubMed] [Google Scholar]
  22. Conti-Ramsden G, Simkin Z, Botting N. The prevalence of autistic spectrum disorders in adolescents with a history of specific language impairment (SLI) Journal of Child Psychology and Psychiatry. 2006;47(6):621–628. doi: 10.1111/j.1469-7610.2005.01584.x. [DOI] [PubMed] [Google Scholar]
  23. Dawson J, Eyer J, Fonkalsrud J. Structured Photographic Expressive Language Test - Preschool: Second Edition. DeKalb, IL: Janelle Publications; 2005. [Google Scholar]
  24. de Guibert C, Maumet C, Jannin P, Ferré J-C, Tréguier C, Barillot C, Biraben A. Abnormal functional lateralization and activity of language brain areas in typical specific language impairment (developmental dysphasia) Brain. 2011;134(Pt 10):3044–3058. doi: 10.1093/brain/awr141. [DOI] [PMC free article] [PubMed] [Google Scholar]
  25. Dennis EL, Thompson PM. Mapping connectivity in the developing brain. International Journal of Developmental Neuroscience. doi: 10.1016/j.ijdevneu.2013.05.007. in press. [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. Di Russo F, Martínez A, Sereno MI, Pitzalis S, Hillyard SA. Cortical sources of the early components of the visual evoked potential. Human Brain Mapping. 2001;15:95–111. doi: 10.1002/hbm.10010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  27. Dick AS, Solodkin A, Small SL. Neural development of networks for audiovisual speech comprehension. Brain and Language. 2010;114:101–114. doi: 10.1016/j.bandl.2009.08.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. Dispaldro M, Leonard LB, Corradi N, Ruffino M, Bronte T, Facoetti A. Visual attentional engagement deficits in children with Specific Language Impairment and their role in real-time language processing. Cortex. doi: 10.1016/j.cortex.2012.09.012. in press. [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Dixon NF, Spitz L. The detection of auditory visual desynchrony. Perception. 1980;9:719–721. doi: 10.1068/p090719. [DOI] [PubMed] [Google Scholar]
  30. Dollaghan C, Campbell TF. Nonword repetition and child language impairment. Journal of Speech, Language, and Hearing Research. 1998;41:1136–1146. doi: 10.1044/jslhr.4105.1136. [DOI] [PubMed] [Google Scholar]
  31. Ellis Weismer S, Tomblin JB, Zhang X, Buckwalter P, Chynoweth JG, Jones M. Nonword repetition performance in school-age children with and without language impairment. Journal of Speech, Language, and Hearing Research. 2000;43:865–878. doi: 10.1044/jslhr.4304.865. [DOI] [PubMed] [Google Scholar]
  32. Finneran D, Francis A, Leonard LB. Sustained attention in children with Specific Language Impairment. Journal of Speech, Language, and Hearing Research. 2009;52(4):915–929. doi: 10.1044/1092-4388(2009/07-0053). [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Flax JF, Realpe-Bonilla T, Hirsch LS, Brzustowicz LM, Bartlett CW, Tallal P. Specific language impairment in families: evidence for co-occurrence with reading impairments. Journal of Speech, Language, and Hearing Research. 2003;46(3):530–543. doi: 10.1044/1092-4388(2003/043). [DOI] [PubMed] [Google Scholar]
  34. Foss-Feig JH, Kwakye LD, Cascio CJ, Burnette CP, Kadivar H, Stone WL, Wallace MT. An extended multisensory temporal binding window in autism spectrum disorders. Experimental Brain Research. 2010;203:381–389. doi: 10.1007/s00221-010-2240-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Graf Estes K, Evans JL, Else-Quest NM. Differences in the nonword repetition performance of children with and without specific langauge impairment: A meta-analysis. Journal of Speech, Language, and Hearing Research. 2007;50:177–195. doi: 10.1044/1092-4388(2007/015). [DOI] [PubMed] [Google Scholar]
  36. Grant KW, van Wassenhove V, Poeppel D. Detection of auditory (cross-spectral) and auditory-visual (cross-modal) synchrony. Speech Communication. 2004;44:43–53. [Google Scholar]
  37. Greenslade K, Plante E, Vance R. The diagnostic accuracy and construct validity of the Structured Photographic Expressive Language Test - Preschool: Second Edition. Language, Speech, and Hearing Services in Schools. 2009;40:150–160. doi: 10.1044/0161-1461(2008/07-0049). [DOI] [PMC free article] [PubMed] [Google Scholar]
  38. Grondin S, Dionne G, Malenfant N, Plourde M, Cloutier M, Jean C. Temporal processing skills of children with and without specific language impairment. Canadian Journal of Speech-Language Pathology and Audiology. 2007;31(1):38–46. [Google Scholar]
  39. Grondin S, Rousseau R. Judging the relative duration of multimodal short empty time intervals. Perception and Psychophysics. 1991;49(3):245–256. doi: 10.3758/bf03214309. [DOI] [PubMed] [Google Scholar]
  40. Hairston WD, Burdette JH, Flowers DL, Wood FB, Wallace MT. Altered temporal profile of visual-auditory multisensory interactions in dyslexia. Experimental Brain Research. 2005;166:474–480. doi: 10.1007/s00221-005-2387-6. [DOI] [PubMed] [Google Scholar]
  41. Hayes EA, Tiippana K, Nicol TG, Sams M, Kraus N. Integration of heard and seen speech: a factor in learning disabilities in children. Neuroscience Letters. 2003;351:46–50. doi: 10.1016/s0304-3940(03)00971-6. [DOI] [PubMed] [Google Scholar]
  42. Hays M. Statistics. 5. Fort Worth, TX: Harcourt Brace College Publishers; 1994. [Google Scholar]
  43. Hesketh A, Conti-Ramsden G. Memory and language in middle childhood in individuals with a history of specific language impairment. PLOS ONE. 2013;8(2) doi: 10.1371/journal.pone.0056314. [DOI] [PMC free article] [PubMed] [Google Scholar]
  44. Hillock-Dunn A, Wallace MT. Developmental changes in the multisensory temporal binding window persist into adolescence. Developmental Science. 2012;15(5):688–696. doi: 10.1111/j.1467-7687.2012.01171.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  45. Hillock AR, Powers AR, Wallace MT. Binding of sights and sounds: Age-related changes in multisensory temporal processing. Neuropsychologia. 2011;49:461–467. doi: 10.1016/j.neuropsychologia.2010.11.041. [DOI] [PMC free article] [PubMed] [Google Scholar]
  46. Hollich G, Newman RS, Jusczyk PW. Infants’ use of synchronized visual information to separate streams of speech. Child Development. 2005;76(3):598–613. doi: 10.1111/j.1467-8624.2005.00866.x. [DOI] [PubMed] [Google Scholar]
  47. Kaganovich N. Development of sensitivity to audivisual temporal asynchrony during mid-childhood in preparation. [Google Scholar]
  48. Karmiloff-Smith A. Nativism versus neuroconstructivism: Rethinking the study of developmental disorders. Developmental Psychology. 2009;45(1):56–63. doi: 10.1037/a0014506. [DOI] [PubMed] [Google Scholar]
  49. Kushnerenko E, Teinonen T, Volein A, Csibra G. Electrophysiological evidence of illusory audiovisual speech percept in human infants. Proceedings of the National Academy of Sciences. 2008;105(32):11442–11445. doi: 10.1073/pnas.0804275105. [DOI] [PMC free article] [PubMed] [Google Scholar]
  50. Kwakye LD, Foss-Feig JH, Cascio CJ, Stone WL, Wallace MT. Altered auditory and multisensory temporal processing in autism spectrum disorders. Frontiers in Integrative Neuroscience. 2011;4(129) doi: 10.3389/fnint.2010.00129. [DOI] [PMC free article] [PubMed] [Google Scholar]
  51. Leonard L. Children with Specific Language Impairment. 2. Cambridge, Massachusetts: The MIT Press; 2014. [Google Scholar]
  52. Leonard L, Davis J, Deevy P. Phonotactic probability and past tense use by children with specific language impairment and their typically developing peers. Clinical Linguistics and Phonetics. 2007;21:747–758. doi: 10.1080/02699200701495473. [DOI] [PMC free article] [PubMed] [Google Scholar]
  53. Leonard L, Deevy P. Input distribution influences degree of auxiliary use by children with SLI. Cognitive Linguistics. 2011;22:247–273. doi: 10.1515/COGL.2011.010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  54. Leonard L, Deevy P, Kurtz R, Krantz Chorev L, Owen A, Polite E, Finneran D. Lexical aspect and the use of verb morphology by children with specific language impairment. Journal of Speech, Language, and Hearing Research. 2007;50:759–777. doi: 10.1044/1092-4388(2007/053). [DOI] [PubMed] [Google Scholar]
  55. Leonard L, Ellis Weismer S, Miller CA, Francis DJ, Tomblin JB, Kail RV. Speed of processing, working memory, and language impairment in children. Journal of Speech, Language, and Hearing Research. 2007;50:408–428. doi: 10.1044/1092-4388(2007/029). [DOI] [PubMed] [Google Scholar]
  56. Leonard L, McGregor KK, Allen GD. Grammatical Morphology and Speech Perception in Children with Specific Language Impairment. Journal of Speech and Hearing Research. 1992;35:1076–1085. doi: 10.1044/jshr.3505.1076. [DOI] [PubMed] [Google Scholar]
  57. Lewkowicz DJ. Perception of auditory-visual temporal synchrony in human infants. Journal of Experimental Psychology: Human Perception and Performance. 1996;22:1094–1106. doi: 10.1037//0096-1523.22.5.1094. [DOI] [PubMed] [Google Scholar]
  58. Lewkowicz DJ. Learning and discrimination of audiovisual events in human infants: The hierarchical relation between intersensory temporal synchrony and rhythmic pattern cues. Developmental Psychology. 2003;39(5):795–804. doi: 10.1037/0012-1649.39.5.795. [DOI] [PubMed] [Google Scholar]
  59. Lewkowicz DJ. Infant perception of audio-visual speech synchrony. Developmental Psychology. 2010;46:66–77. doi: 10.1037/a0015579. [DOI] [PubMed] [Google Scholar]
  60. Lewkowicz DJ, Hansen-Tift AM. Infants deploy selective attention to the mouth of a talking face when learning speech. Proceedings of the National Academy of Sciences. 2012;109(5):1431–1436. doi: 10.1073/pnas.1114783109. [DOI] [PMC free article] [PubMed] [Google Scholar]
  61. Luck SJ. An Introduction to the Event-Related Potential Technique. Cambridge, MA: The MIT Press; 2005. [Google Scholar]
  62. Manly T, Robertson IH, Anderson V, Nimmo-Smith I. Test of Everyday Attention for Children. London, UK: Pearson Assessment; 1999. [DOI] [PubMed] [Google Scholar]
  63. Martin N, Brownel R. Test of Auditory Processing Skills. 3. Novato, California: Academic Therapy Publications; 2005. [Google Scholar]
  64. Massaro DW. Children’s perception of visual and auditory speech. Child Development. 1984;55:1777–1788. [PubMed] [Google Scholar]
  65. Massaro DW, Thompson LA, Barron B, Lauren E. Developmental changes in visual and auditory contributions to speech perception. Journal of Experimental Child Psychology. 1986;41:93–113. doi: 10.1016/0022-0965(86)90053-6. [DOI] [PubMed] [Google Scholar]
  66. McArthur GM, Bishop DVM. Speech and non-speech processing in people with specific language impairment: A behavioral and electrophysiological study. Brain and Language. 2005;94:260–273. doi: 10.1016/j.bandl.2005.01.002. [DOI] [PubMed] [Google Scholar]
  67. McCulloch DL. Visual evoked potentials in infants. In: de Haan M, editor. Infant EEG and Event-Related Potentials. New York: Psychology Press; 2007. pp. 39–64. [Google Scholar]
  68. McGrath M, Summerfield Q. Intermodal timing relations and audio-visual speech recognition by normal-hearing adults. Journal of the Acoustical Society of America. 1985;77(2):678–685. doi: 10.1121/1.392336. [DOI] [PubMed] [Google Scholar]
  69. McGurk H, MacDonald J. Hearing lips and seeing voices. Nature. 1976;264:746–748. doi: 10.1038/264746a0. [DOI] [PubMed] [Google Scholar]
  70. Meredith MA, Nemitz JW, Stein BE. Determinants of multisensory integration in superior colliculus neurons. I. Temporal factors. The Journal of Neuroscience. 1987;7(10):3215–3229. doi: 10.1523/JNEUROSCI.07-10-03215.1987. [DOI] [PMC free article] [PubMed] [Google Scholar]
  71. Meronen A, Tiippana K, Westerholm J, Ahonen T. Audiovisual speech perception in children with developmental language disorder in degraded listening conditions. Journal of Speech, Language, and Hearing Research. 2013;56:211–221. doi: 10.1044/1092-4388(2012/11-0270). [DOI] [PubMed] [Google Scholar]
  72. Metting van Rijn AC, Kuiper AP, Dankers TE, Grimbergen CA. Low-cost active electrode improves the resolution in biopotential recordings. Paper presented at the 18th Annual International Conference of the IEEE Engineering in Medicine and Biology Society; Amsterdam, The Netherlands. 1996. [Google Scholar]
  73. Metting van Rijn AC, Peper A, Grimbergen CA. High-quality recording of bioelectric events. Part 1: Interference reduction, theory and practice. Medical and Biological Engineering and Computing. 1990;28:389–397. doi: 10.1007/BF02441961. [DOI] [PubMed] [Google Scholar]
  74. Näätänen R, Picton T. The N1 wave of the human electric and magnetic response to sound: A review and an analysis of the component structure. Psychophysiology. 1987;24(4):375–425. doi: 10.1111/j.1469-8986.1987.tb00311.x. [DOI] [PubMed] [Google Scholar]
  75. Neville HJ, Coffey SA, Holcomb PJ, Tallal P. The neurobiology of sensory and language processing in language-impaired children. Journal of Cognitive Neuroscience. 1993;5(2):235–253. doi: 10.1162/jocn.1993.5.2.235. [DOI] [PubMed] [Google Scholar]
  76. Norrix LW, Plante E, Vance R. Auditory-visual speech integration by adults with and without language-learning disabilities. Journal of Communication Disorders. 2006;39:22–36. doi: 10.1016/j.jcomdis.2005.05.003. [DOI] [PubMed] [Google Scholar]
  77. Norrix LW, Plante E, Vance R, Boliek CA. Auditory-visual integration for speech by children with and without specific language impairment. Journal of Speech, Language, and Hearing Research. 2007;50:1639–1651. doi: 10.1044/1092-4388(2007/111). [DOI] [PubMed] [Google Scholar]
  78. Oldfield RC. The assessment and analysis of handedness: The Edinburgh inventory. Neuropsychologia. 1971;9:97–113. doi: 10.1016/0028-3932(71)90067-4. [DOI] [PubMed] [Google Scholar]
  79. Pang EW, Taylor MJ. Tracking the development of the N1 from age 3 to adulthood: an examination of speech and non-speech stimuli. Clinical Neurophysiology. 2000;111:388–397. doi: 10.1016/s1388-2457(99)00259-x. [DOI] [PubMed] [Google Scholar]
  80. Petrini K, Dahl S, Rocchesso D, Waadeland CH, Avanzini D, Puce A, Pollick FE. Multisensory integration of drumming actions: musical expertise affects perceived audiovisual asynchrony. Experimental Brain Research. 2009;198:339–352. doi: 10.1007/s00221-009-1817-2. [DOI] [PubMed] [Google Scholar]
  81. Petrini K, Halt SP, Pollick F. Expertise with multisensory events eliminates the effect of biological motion rotation on audiovisual synchrony perception. Journal of Vision. 2010;10(5 2):1–14. doi: 10.1167/10.5.2. [DOI] [PubMed] [Google Scholar]
  82. Petrini K, Russell M, Pollick F. When knowing can replace seeing in audiovisual integration of actions. Cognition. 2009;110:432–439. doi: 10.1016/j.cognition.2008.11.015. [DOI] [PubMed] [Google Scholar]
  83. Pflieger ME. Theory of a spatial filter for removing ocular artifacts with preservation of EEG; Paper presented at the EMSE Workshop; Princeton University; 2001. http://www.sourcesignal.com/SpFilt_Ocular_Artifact.pdf. [Google Scholar]
  84. Plante E, Vance R. Selection of preschool language tests: A data-based approach. Language, Speech, and Hearing Services in the Schools. 1994;25:15–24. [Google Scholar]
  85. Pons F, Andreu L, Sanz-Torrent M, Buil-Legaz L, Lewkowicz DJ. Perception of audiovisual speech synchrony in Spanish-speaking children with and without specific language impairment. Journal of Child Language. 2013;40(3):687–700. doi: 10.1017/S0305000912000189. [DOI] [PMC free article] [PubMed] [Google Scholar]
  86. Pons F, Lewkowicz DJ, Soto-Faraco S, Sebastián-Gallés N. Narrowing of intersensory speech perception in infancy. Proceedings of the National Academy of Sciences. 2009;106(26):10598–10602. doi: 10.1073/pnas.0904134106. [DOI] [PMC free article] [PubMed] [Google Scholar]
  87. Pratt H. Sensory ERP components. In: Luck SJ, Kappenman ES, editors. The Oxford Handbook of Event-Related Potential Components. Oxford: Oxford University Press; 2011. pp. 89–114. [Google Scholar]
  88. Redmond SM. Differentiating SLI from ADHD using children’s sentence recall and production of past tense morphology. Clinical Linguistics and Phonetics. 2005;19(2):109–127. doi: 10.1080/02699200410001669870. [DOI] [PubMed] [Google Scholar]
  89. Redmond SM, Thompson HL, Goldstein S. Psycholinguistic profiling differentiates Specific Language Impairment from typical development and from Attention-Deficit/Hyperactivity Disorder. Journal of Speech, Language, and Hearing Research. 2011;54:99–117. doi: 10.1044/1092-4388(2010/10-0010). [DOI] [PMC free article] [PubMed] [Google Scholar]
  90. Rice ML, Hoffman L, Wexler K. Judgments of omitted BE and DO in questions as extended finiteness clinical markers of Specific Language Impairment (SLI) to 15 years: A study of growth and asymptote. Journal of Speech, Language, and Hearing Research. 2009;52:1417–1433. doi: 10.1044/1092-4388(2009/08-0171). [DOI] [PMC free article] [PubMed] [Google Scholar]
  91. Rosenblum LD, Schmuckler MA, Johnson JA. The McGurk effect in infants. Perception and Psychophysics. 1997;59(3):347–357. doi: 10.3758/bf03211902. [DOI] [PubMed] [Google Scholar]
  92. Scarborough HS, Dobrich W. Development of children with early language delay. Journal of Speech and Hearing Research. 1990;33:70–83. doi: 10.1044/jshr.3301.70. [DOI] [PubMed] [Google Scholar]
  93. Scherg M, Vajsar J, Picton T. A source analysis of the late human auditory evoked potentials. Journal of Cognitive Neuroscience. 1989;1(4):336–355. doi: 10.1162/jocn.1989.1.4.336. [DOI] [PubMed] [Google Scholar]
  94. Schopler E, Van Bourgondien ME, Wellman GJ, Love SR. Childhood Autism Rating Scale. 2. Western Psychological Services; 2010. [Google Scholar]
  95. Schorr EA, Fox NA, van Wassenhove V, Knudsen EI. Auditory-visual fusion in speech perception in children with cochlear implants. Proceedings of the National Academy of Sciences. 2005;102(51):18748–18750. doi: 10.1073/pnas.0508862102. [DOI] [PMC free article] [PubMed] [Google Scholar]
  96. Semel E, Wiig EH, Secord WA. CELF4: Clinical Evaluation of Language Fundamentals. 4. San Antonio, TX: Pearson Clinical Assessment; 2003. [Google Scholar]
  97. Sharma A, Kraus N, McGee TJ, Nicol TG. Developmental changes in P1 and N1 central auditory responses elicited by consonant-vowel syllables. Electroencephalography and Clinical Neurophysiology. 1997;104:540–545. doi: 10.1016/s0168-5597(97)00050-6. [DOI] [PubMed] [Google Scholar]
  98. Spaulding TJ, Plante E, Vance R. Sustained selective attention skills of preschool children with specific language impairment: Evidence for separate attentional capacities. Journal of Speech, Language, and Hearing Research. 2008;51:16–34. doi: 10.1044/1092-4388(2008/002). [DOI] [PubMed] [Google Scholar]
  99. Stekelenburg JJ, Vroomen J. An event-related potential investigation of the time-course of temporal ventriloquism. NeuroReport. 2005;(16):641–644. doi: 10.1097/00001756-200504250-00025. [DOI] [PubMed] [Google Scholar]
  100. Stevens C, Paulsen D, Yasen A, Mitsunaga L, Neville HJ. Electrophysiological evidence for attenuated auditory recovery cycles in children with specific language impairment. Brain Research. 2012;1438:35–47. doi: 10.1016/j.brainres.2011.12.039. [DOI] [PMC free article] [PubMed] [Google Scholar]
  101. Stevens C, Sanders L, Neville H. Neurophysiological evidence for selective auditory attention deficits in children with specific language impairment. Brain Research. 2006;1111(1):143–152. doi: 10.1016/j.brainres.2006.06.114. [DOI] [PubMed] [Google Scholar]
  102. Stevenson RA, Siemann JK, Schneider BC, Eberly HE, Woynaroski TG, Camarata SM, Wallace MT. Multisensory temporal integration in autism spectrum disorders. The Journal of Neuroscience. 2014;34(3):691–697. doi: 10.1523/JNEUROSCI.3615-13.2014. [DOI] [PMC free article] [PubMed] [Google Scholar]
  103. Stevenson RA, Zemtsov RK, Wallace MT. Individual differences in the multisensory temporal binding window predict susceptibility to audiovisual illusions. Journal of Experimental Psychology: Human Perception and Performance. 2012;38(6):1517–1529. doi: 10.1037/a0027339. [DOI] [PMC free article] [PubMed] [Google Scholar]
  104. Sumby WH, Pollack I. Visual contribution to speech intelligibility in noise. Journal of the Acoustical Society of America. 1954;26(2):212–215. [Google Scholar]
  105. Sussman E, Steinschneider M, Gumenyuk V, Grushko J, Lawson K. The maturation of human evoked brain potentials to sounds presented at different stimulus rates. Hearing Research. 2008;236:61–79. doi: 10.1016/j.heares.2007.12.001. [DOI] [PMC free article] [PubMed] [Google Scholar]
  106. Tallal P. Perceptual requisites for language. In: Schiefelbusch R, editor. Non-Speech Language and Communication. Baltimore: University Park Press; 1980. pp. 449–467. [Google Scholar]
  107. Tallal P, Piercy M. Deficits of non-verbal auditory perception in children with developmental aphasia. Nature. 1973a;241:468–469. doi: 10.1038/241468a0. [DOI] [PubMed] [Google Scholar]
  108. Tallal P, Piercy M. Developmental aphasia: Impaired rate of non-verbal processing as a function of sensory modality. Neuropsychologia. 1973b;11:389–398. doi: 10.1016/0028-3932(73)90025-0. [DOI] [PubMed] [Google Scholar]
  109. Tallal P, Stark RE, Mellits DE. Identification of language-impaired children on the basis of rapid perception and production skills. Brain and Language. 1985;25:314–322. doi: 10.1016/0093-934x(85)90087-2. [DOI] [PubMed] [Google Scholar]
  110. Teinonen T, Aslin RN, Alku P, Csibra G. Visual speech contributes to phonetic learning in 6-month-old infants. Cognition. 2008;108:850–855. doi: 10.1016/j.cognition.2008.05.009. [DOI] [PubMed] [Google Scholar]
  111. Van Wassenhove V, Grant KW, Poeppel D. Temporal window of integration in auditory-visual speech perception. Neuropsychologia. 2007;45:598–607. doi: 10.1016/j.neuropsychologia.2006.01.001. [DOI] [PubMed] [Google Scholar]
  112. Vatakis A, Spence C. Audiovisual synchrony perception for music, speech, and object actions. Brain Research. 2006;1111:134–142. doi: 10.1016/j.brainres.2006.05.078. [DOI] [PubMed] [Google Scholar]
  113. Vroomen J, Keetels M. Perception of intersensory synchrony: A tutorial review. Attention, Perception, & Psychophysics. 2010;72(4):871–884. doi: 10.3758/APP.72.4.871. [DOI] [PubMed] [Google Scholar]
  114. Vroomen J, Stekelenburg JJ. Visual anticipatory information modulates multisensory interactions of artificial audiovisual stimuli. Journal of Cognitive Neuroscience. 2009;22(7):1583–1596. doi: 10.1162/jocn.2009.21308. [DOI] [PubMed] [Google Scholar]
  115. Vroomen J, Stekelenburg JJ. Perception of intersensory synchrony in audiovisual speech: Not that special. Cognition. 2011;118(1):75–83. doi: 10.1016/j.cognition.2010.10.002. [DOI] [PubMed] [Google Scholar]
  116. Weber-Fox C, Leonard L, Hampton Wray A, Tomblin JB. Electrophysiological correlates of rapid auditory and linguistic processing in adolescents with specific language impairment. Brain and Language. 2010;115:162–181. doi: 10.1016/j.bandl.2010.09.001. [DOI] [PMC free article] [PubMed] [Google Scholar]
  117. Weikum WM, Vouloumanos A, Navarra J, Soto-Faraco S, Sebastián-Gallés N, Werker JF. Visual language discrimination in infancy. Science. 2007;316:1159. doi: 10.1126/science.1137686. [DOI] [PubMed] [Google Scholar]
  118. Werner EO, Kresheck JD. Structured Photographic Expressive Language Test - II. DeKalb, IL: Janelle Publications; 1983. [Google Scholar]
  119. Whitehouse AJO, Bishop DVM. Cerebral dominance for language function in adults with specific language impairment or autism. Brain. 2008;131:3193–3200. doi: 10.1093/brain/awn266. [DOI] [PMC free article] [PubMed] [Google Scholar]
  120. Yi A, Wong W, Eizenman M. Gaze patterns and audiovisual speech enhancement. Journal of Speech, Language, and Hearing Research. 2013;56:471–480. doi: 10.1044/1092-4388(2012/10-0288). [DOI] [PubMed] [Google Scholar]

RESOURCES