Significance
Language comprehension is thought to rely on a combination of specialized and general-purpose neural mechanisms. When people listen to speech, low-frequency oscillations in cerebral cortex (8 Hz) become entrained to quasirhythmic fluctuations in volume. Entrainment of auditory cortex to speech rhythms has been well documented, but its functional significance has remained unclear. By showing similar entrainment of visual cortex to sign language, we establish that this phase-locking is not due to specific properties of auditory cortex or of oral speech perception. Rather, low-frequency entrainment is a generalized cortical strategy for boosting perceptual sensitivity to informational peaks in time-varying signals.
Keywords: sign language, cortical entrainment, oscillations, EEG
Abstract
Despite immense variability across languages, people can learn to understand any human language, spoken or signed. What neural mechanisms allow people to comprehend language across sensory modalities? When people listen to speech, electrophysiological oscillations in auditory cortex entrain to slow (8 Hz) fluctuations in the acoustic envelope. Entrainment to the speech envelope may reflect mechanisms specialized for auditory perception. Alternatively, flexible entrainment may be a general-purpose cortical mechanism that optimizes sensitivity to rhythmic information regardless of modality. Here, we test these proposals by examining cortical coherence to visual information in sign language. First, we develop a metric to quantify visual change over time. We find quasiperiodic fluctuations in sign language, characterized by lower frequencies than fluctuations in speech. Next, we test for entrainment of neural oscillations to visual change in sign language, using electroencephalography (EEG) in fluent speakers of American Sign Language (ASL) as they watch videos in ASL. We find significant cortical entrainment to visual oscillations in sign language <5 Hz, peaking at 1 Hz. Coherence to sign is strongest over occipital and parietal cortex, in contrast to speech, where coherence is strongest over the auditory cortex. Nonsigners also show coherence to sign language, but entrainment at frontal sites is reduced relative to fluent signers. These results demonstrate that flexible cortical entrainment to language does not depend on neural processes that are specific to auditory speech perception. Low-frequency oscillatory entrainment may reflect a general cortical mechanism that maximizes sensitivity to informational peaks in time-varying signals.
Languages differ dramatically from one another, yet people can learn to understand any natural language. What neural mechanisms allow humans to understand the vast diversity of languages and to distinguish linguistic signal from noise? One mechanism that has been implicated in language comprehension is neural entrainment to the volume envelope of speech. The volume envelope of speech fluctuates at low frequencies ( Hz), decreasing at boundaries between syllables, words, and phrases. When people listen to speech, neural oscillations in the delta (1–4 Hz) and theta (4–8 Hz) bands become entrained to these fluctuations in volume (1–4).
Entrainment to the volume envelope may represent an active neural mechanism to boost perceptual sensitivity to rhythmic stimuli (2, 5–7). Although entrainment is partly driven by bottom-up features of the stimulus (8–10), it also depends on top-down signals to auditory cortex from other brain areas. Auditory entrainment is strengthened when people see congruent visual and auditory information (11, 12) and is modulated by attention (13) and by top-down signals from frontal cortex (4, 14).
Cortical entrainment is proposed to perform a key role in speech comprehension, such as segmenting out syllables from a continuous speech stream (1, 2, 15) or optimizing perceptual sensitivity to rhythmic pulses of sound (5–7). However, the mechanisms driving entrainment to speech remain unclear. We consider two hypotheses. First, flexible entrainment to quasiperiodic rhythms may be specific to auditory perception (6); in visual perception, by contrast, cortical oscillations in the alpha band (8–12 Hz) may phase-lock only to consistent stimulus rhythms, without adjusting to variable stimulus rhythms (16). Second, low-frequency cortical entrainment may be a general-purpose neural mechanism that helps optimize perception to time-varying stimuli regardless of the perceptual modality. Neural oscillations may allow the brain to rhythmically orient attention to quasiperiodic stimuli (5, 17) across sensory systems.
Because previous studies of cortical entrainment to rhythms in language have focused on oral speech, they have been unable to distinguish between these competing hypotheses. Here, we test for low-frequency entrainment to a purely visual language: American Sign Language (ASL). Prior studies have shown that neural and behavioral oscillations in vision are preferentially entrained by stimuli that flicker in the alpha band (18–21). Therefore, if flexible cortical entrainment to oral speech depends on modality-specific properties of auditory processing, then phase-locking to sign language should be concentrated in the alpha band, if it occurs at all (16). Alternatively, if cortical entrainment is a generalized neural strategy to maximize sensitivity to rhythmic stimuli, then oscillatory activity in visual cortex should entrain at the frequency of informational changes in ASL.
To determine whether human cerebral cortex entrains to rhythmic information in sign language, first we developed a metric for quantifying the amplitude of visual change in sign, by analogy to the acoustic envelope of speech. Next, we characterized visual variability across four sign languages, showing that this variability is quasiperiodic at <8 Hz. Finally, we demonstrated that cerebral cortex entrains to visual variability in sign language and showed that entrainment is strongest around the frequencies of phrases and individual signs in ASL.
Results
Developing a Metric for Quantifying Visual Change.
To examine neural entrainment to visual rhythms in sign language, we must first quantify the “amplitude envelope” of a visual signal. The acoustic envelope is a highly reduced representation of sound, tracing extreme amplitude values in the time-varying signal. Oscillations in the envelope of speech depend on movements of various components of the vocal tract, including the rhythmic opening and closing of the mandible (22). Sign language, in contrast, does not involve consistent oscillatory movements by any single effector (23). However, quasiperiodic oscillations in sign language may arise from the coordinated movements of multiple effectors.
Here we present the Instantaneous Visual Change (IVC) as a metric that is conceptually similar to the acoustic envelope, summarizing the amplitude of change at each time point. The IVC is a time series of aggregated visual changes between frames (Fig. 1 and Method and Materials). This algorithm provides an automatic, objective alternative to human-coded methods of studying temporal structure in sign (24).
Fig. 1.
Calculation of the IVC. The IVC summarizes total visual change at each point in time. First, the difference between adjacent grayscale video frames (Top) is calculated for each pixel. To aggregate over both increases and decreases in brightness, these pixel-wise differences are then squared (Middle). Finally, the brightness values in all pixels of the squared-difference images are summed to obtain a single value summarizing the magnitude of change between two video frames. Computation of this value for each adjacent pair of frames yields a time-series (Bottom).
The amplitude of the IVC indexes the amount of visual change between two video frames. The largest peaks in the IVC therefore occur during large, quick movements. In the videos we analyzed, these changes corresponded primarily to movements of the signers’ hands and arms, but may also reflect movements of the face, head, and torso. For example, a quick arm movement results in a larger number of pixels changing in each adjacent frame—and a higher peak in the IVC—than a slow arm movement. The IVC thus offers a heuristic index of linguistic information in the visual signal. An example movie illustrating the IVC is included as Movie S1.
Characterizing Temporal Structure in Sign Language.
The IVC allows us to characterize one dimension of the temporal structure in sign language and to directly compare the spectral signatures of amplitude variability across sign and oral speech. Visual examination of the raw IVC of sign language reveals quasiperiodic oscillations with irregularly timed peaks (Fig. 2A). To characterize variability within and across sign languages, we computed the power spectra of the IVC from samples of four different sign languages: American Sign Language (ASL), German Sign Language (Deutsche Gebärdensprache; DGS), British Sign Language (BSL), and Australian Sign Language (Auslan). These languages developed independently of the oral languages spoken in these countries and come from three genetically unrelated language families (BSL and Auslan from the British, Australian, and New Zealand Sign Language family; ASL from the French Sign Language family; and DGS from the German Sign Language family). In all four languages, power in the IVC decreases roughly monotonically with increasing frequency, without any pronounced peaks in the spectrum (Fig. 2C). We tested for rhythmic components in the IVC by comparing these spectra against the spectrum characteristic of many signals in the natural world. Power that is higher than the function indicates periodicity at that frequency (22). The IVC of sign language showed elevated power at ∼2–8 Hz (). Individual signs in sign languages tend to occur at ∼2–2.5 Hz (25, 26), on the lower end of the rhythmic components in the IVC. These analyses suggest that sign language involves weak, quasiperiodic rhythms with variable frequencies in the delta and theta range.
Fig. 2.
Temporal structure in signed and spoken language. (A) Example trace of the IVC of ASL. (B) Example trace of the broadband envelope (env.) of English. (C) Spectrum of sign language plotted on log–log axes. Line color denotes language. Each curve shows the spectrum of a separate video sample (; total duration 1:10:22). The black line shows the best-fit trend across all samples. The gray bar at the base of the plot shows where the IVC spectra are significantly greater than the fit ( by one-sample t tests). (D) Comparison of the mean spectra across signed and spoken languages. Shaded area depicts the SE of the mean. The gray bar indicates significant differences between the two curves ( by independent-samples t tests). Sign language samples are the same as in C. Audio recordings were sampled from speech in nine languages (; total duration 1:07:28). Amplitude in all analyses has been standardized by each recording’s SD.
To explicitly compare temporal structure between sign and speech, we contrasted the IVC of sign with the broadband envelope of speech (22). We computed the broadband envelope of samples from nine spoken languages representing five language families: English, French, Portuguese, Dutch, German, Hungarian, Japanese, Arabic, and Mandarin. After resampling the broadband envelopes and IVC signals to a common frequency (30 Hz) and standardizing the amplitude of each recording by dividing out its standard deviation, we compared the average spectra of the IVC of sign and the broadband envelope of speech (Fig. 2D). Spoken languages showed stronger modulations than sign languages at >2 Hz. This increased power may reflect modulation due to syllables in speech, which occur at ∼2–10 Hz (22, 27). Indeed, peaks from individual syllables are visible in the broadband envelope, and these peaks occur at 4 Hz (Fig. 2B). These results indicate that visual motion in sign language is modulated at lower frequencies than auditory volume in spoken language. This difference is consistent with the slower movements in the articulators for sign (the hands) than in the articulators for speech (the vocal tract) (25, 26).
Cortical Coherence to Visual Rhythms in Sign Language.
We used electroencephalography (EEG) to examine cortical entrainment to quasirhythmic fluctuations in visual information in sign language. Fluent speakers of ASL watched videos of ASL stories against a static background. We tested for coherence of low-frequency electrophysiological oscillations to quasiperiodic oscillations in the IVC.
Coherence was calculated separately at each EEG channel in partially overlapping, logarithmically spaced bins centered over 0.5–16 Hz (Materials and Methods). Because coherence provides no intrinsic measure of chance performance, we created a null distribution of coherence using a randomization procedure. To obtain each value in the null distribution, we time-shifted the IVC to a randomly selected starting point, moving the portion of the IVC that remained after the final time point of the EEG signal to the beginning of the recording. This procedure preserves the spectral and temporal characteristics of the EEG and IVC recordings, but eliminates any relationship between these signals. Coherence was then computed between the EEG recordings and the randomly shifted IVC.
A cluster-based permutation test indicated that coherence between cortical oscillations and the IVC of sign was stronger than would be expected by chance (). Averaging the coherence spectrum across every EEG channel, coherence was above chance from 0.4 to 5 Hz, peaking at 1 Hz (Fig. 3A). Coherence emerged over a similar range of frequencies when we selected only occipital channels (0.8–5 Hz; ), primarily reflecting entrainment in visual cortex (Fig. 3B). In frontal channels, above-chance coherence was present from 0.4 to 1.25 Hz (; Fig. 3C), revealing top-down control from frontal cortex. Examining the entire scalp distribution, cortical coherence to the IVC of sign language was strongest over central and occipital channels (Fig. 3D).
Fig. 3.
Coherence between cortex and the IVC of sign language in fluent signers (). (A) Coherence spectrum averaged across all EEG channels. For each participant, we computed the difference in empirical coherence and a distribution of cortical coherence to randomly shifted IVC. The solid line shows the mean difference in empirical and randomized coherence across participants, and the shaded area shows the 95% CI around the mean. The dashed line shows chance levels. (B) Coherence spectrum averaged over occipital channels. Inset shows the location of selected channels. (C) Coherence spectrum averaged over frontal channels. Inset shows the location of selected channels. (D) Scalp topography of coherence in each frequency bin, averaged across participants. Diff., difference.
To test whether cortical entrainment depends on linguistic knowledge, we examined coherence to sign language in people who did not know any ASL. Like signers, nonsigners showed significant coherence to videos of ASL storytelling (), with the strongest coherence over central and occipital channels from 0.8 to 3.5 Hz (Fig. S1). We then separately analyzed effects of linguistic knowledge on entrainment in occipital and frontal cortex. Although coherence at occipital channels did not significantly differ between groups (Fig. 4B), coherence at frontal channels was stronger in signers than in nonsigners, indicating differences in top-down control based on familiarity with ASL (; Fig. 4A and Fig. S1).
Fig. S1.
Comparison of coherence across signing and nonsigning participants. Coherence was robustly present in both groups of participants, regardless of experience with sign language. The top row shows coherence in each frequency bin for fluent signers (same data as in Fig. 3), and the middle row shows coherence for nonsigners. The bottom row shows the difference in coherence between the two groups. Positive values (shown in red) indicate stronger coherence in signers, and negative values (blue) indicate stronger coherence in nonsigners.
Fig. 4.
Comparison of coherence in fluent signers () and nonsigners (). (A) Coherence at frontal channels (channel selection illustrated at right) was stronger in signers than in nonsigners. (B) Coherence at occipital channels did not differ between groups. Data from signers is the same as in Fig. 3. *; n.s., not statistically significant. Diff., difference.
Discussion
Cortical Coherence to Sign Language.
In this study, we find that electrophysiological oscillations in human cerebral cortex become entrained to quasiperiodic fluctuations of visual movement in sign language. In fluent signers, cortical entrainment to sign language was found between 0.4 and 5 Hz, peaking at 1 Hz, and emerged most robustly over occipital and central EEG channels. These results show that the human brain entrains to low-frequency variability in language whether it is perceived with the ears or eyes. Visual cortex flexibly phase-locks to visible changes in sign language, analogously to the way auditory cortex phase-locks to amplitude changes in oral speech. Our findings argue that flexible entrainment depends on mechanisms that are not specific to any given effector or sensory modality.
Prior results suggest that auditory and visual perception are differentially modulated by rhythms at different frequencies (16). Auditory sensitivity varies as a function of the power and phase of spontaneous 2- to 6-Hz rhythms (28), and these oscillations are entrained by sounds modulated at 3 Hz (29, 30). Visual sensitivity, by contrast, depends on the power and phase of spontaneous alpha rhythms (31–33), and electrophysiological oscillations in visual cortex are robustly entrained by periodic stimulation at 10 Hz (19, 21). When humans watch a light flicker at frequencies from 1 to 100 Hz, visual cortex shows the strongest entrainment at 10 Hz (18).
Although these rhythmic preferences are not absolute [visual cortex also shows rhythmic oscillations in the delta/theta range (11, 17, 34–37)], differences between the sensory modalities are apparent when different frequency bands are directly compared. Auditory detection sensitivity depends on the phase of underlying delta–theta, but not alpha, oscillations (28). In response to aperiodic stimulation, visual cortex oscillates in the alpha band (20), whereas auditory cortex does not show consistent oscillatory activity (38).
If sensory-specific oscillatory preferences determine the spectrum of entrainment, then peak coherence to sign language should be observed at the frequencies preferred by visual cortex: 10 Hz. Contrary to this prediction, we find that cortical coherence to sign language only emerges at <5 Hz. Cerebral cortex entrains to sign language around the frequencies of words and phrases in ASL.
Coherence Across Signers and Nonsigners.
We find that cerebral cortex phase-locks to visual changes in ASL both in fluent signers and in people with no knowledge of sign language. In principle, coherence to ASL in nonsigners could emerge for two reasons. Coherence could be driven either bottom-up, by sensory stimulation, or top-down, by nonlinguistic temporal predictions. Human bodies move in predictable ways, and nonsigners could entrain to sign language based on these regularities in human movement.
In frontal areas, fluent signers showed stronger coherence than nonsigners. This difference in frontal coherence may reflect top-down sensory predictions based on knowledge of ASL. Alternatively, differences in coherence could reflect differences in attention to the videos. Cortical entrainment to oral speech decreases when people direct attention away from the speech stimulus (13, 39). Reduced coherence in nonsigners, therefore, would also be predicted if nonsigners do not attend to videos of ASL as strongly as fluent signers do. However, our findings at occipital channels argue against this possibility. If differences between groups were driven by attention, then occipital coherence should be stronger in signers than in nonsigners. However, we find no evidence that occipital coherence depends on linguistic knowledge. Together, these results suggest that, although linguistic knowledge is not necessary for entrainment, signers may leverage knowledge about ASL to sharpen temporal predictions during language comprehension. These sharpened predictions result in stronger entrainment in the frontal regions that exert top-down control over visual perception.
Specialization for Speech?
Syllables in oral speech occur at frequencies that largely overlap with cortical entrainment to the volume envelope. This overlap could be interpreted as evidence for a specialized oscillatory mechanism for speech comprehension. This type of speech-specific mechanism could evolve in at least two ways. First, as Giraud and Poeppel state: “The articulatory motor system [may have] structured its output to match those rhythms the auditory system can best apprehend” (6). Second, auditory mechanisms may have developed to comprehend speech based on the timing of preexisting oral behaviors. Nonhuman primates create vocalizations and facial displays that fluctuate at frequencies similar to human speech syllables (40), and their attention is preferentially captured by faces that move at these frequencies (41); perhaps auditory processing evolved to fit the timing profile of these behaviors.
The data we report here, however, suggest that entrainment may not have any close evolutionary link to oral speech. Instead, a more general process may underlie cortical phase-locking to variability in language. Previous results are consistent with this interpretation as well. When participants watch videos of speech, entrainment emerges not only in auditory cortex, but also in visual cortex (11, 35, 36). Furthermore, cortical rhythms entrain to rhythms in music (42) and to other rhythmic stimuli in audition (29, 30) and vision (19, 43). These examples of low-frequency cortical entrainment to a broad range of stimuli across sensory modalities suggest that the cortical mechanisms supporting entrainment to the volume envelope of speech may be a specialized case of a general predictive process.
Neural Mechanisms of Language Comprehension Across Sensory Modalities.
Previous studies have shown that the functional neuroanatomy of speech largely overlaps with that of sign (44). At the coarsest level of anatomical specificity, the left hemisphere is specialized for spoken language. The left hemisphere is also asymmetrically active during sign language perception (45) and production, regardless of which hand people use to sign (46, 47). Left hemisphere damage, furthermore, results in linguistic deficits in signing patients (48).
Specific regions within the left hemisphere show similar involvement in processing both speech and sign. Across signed and spoken language, bloodflow increases to the left inferior frontal gyrus (LIFG) and left inferior parietal lobe (IPL) during phonemic discrimination (49, 50) and morphosyntactic processing (51). Similarly, word production in both signed and spoken languages activates LIFG, left IPL, and left temporal areas (52).
Differences in the cortical areas involved in sign and speech can often be attributed to differences in the form of these languages. For example, comprehension of sign language activates primary visual, but not primary auditory, cortex (45). Consistent with the fact that sign language relies on spatial contrasts, inferior and superior parietal cortex is more strongly active during signed than during spoken language production (52) and perception (49).
Our findings go beyond functional neuroanatomy to examine neurophysiological processes that can arise in multiple cortical areas. We show that oscillatory entrainment to low-frequency variability in the stimulus occurs, regardless of whether language is being processed using auditory cortex or visual cortex.
Our results differ from previous studies on entrainment to speech primarily in the scalp topography of coherence. Entrainment to auditory speech is strongest over auditory cortex (2, 8, 11, 35) and central frontal sites (14). By contrast, our results show that entrainment to sign language is strongest at occipital and parietal channels, consistent with greater parietal activation during sign compared with speech (53). This difference likely reflects increased visual and spatial demands of perceiving sign language.
The IVC Quantifies Temporal Structure in Visual Perception.
The IVC provides a method for examining gross temporal structure in natural visual stimuli. Analogously to the way the broadband envelope summarizes early stages of auditory processing, the IVC provides a first approximation of the magnitude of information available to the earliest stages of visual processing. At the first stage of auditory transduction, hair cells in the cochlea extract the narrowband envelope of sounds. Summing these narrowband envelopes together yields the overall auditory responses over time: the broadband envelope. In the retina, center-surround retinal ganglion cells respond to changes in the brightness of specific wavelengths of light. Summing the responses from these cells yields the overall visual responses over time, approximated by the IVC.
The IVC provides a coarse index of visual information in sign language, just as the broadband envelope provides a coarse index of information in speech. For example, the volume envelope does not reflect small spectral differences that are crucial for discriminating vowels. The IVC, analogously, does not preserve information about which effectors are moving or their trajectories. Nevertheless, sign language comprehenders may use the IVC of sign heuristically, as listeners use the acoustic envelope of speech, to anticipate when important information is likely to appear.
In the present study, we use the IVC to characterize temporal structure in sign language and to examine responses of the human brain to that temporal structure. The IVC could also be applied to study temporal structure in other domains, such as gesture, biological motion, and movement in natural scenes.
The Functional Role of Entrainment to Language.
Oscillatory entrainment to language may be a specific case of a general cortical mechanism. In primates, spiking probability varies with the phase of low-frequency oscillations: Neurons are most likely to fire at specific points in the phase of ongoing oscillations (54, 55). Perhaps the cortex strategically resets the phase of ongoing neural oscillations to ensure that perceptual neurons are in an excitable state when new information is likely to appear (5–7, 14). Oscillatory entrainment may constitute a cortical strategy to boost perceptual sensitivity at informational peaks in language. Our findings suggest that the brain can flexibly entrain to linguistic information regardless of the modality in which language is produced or perceived.
Materials and Methods
Participants watched ∼20 min of naturalistic storytelling in ASL while EEG was recorded. Participants were instructed to watch the videos and remain still and relaxed. All procedures were approved by the Institutional Review Board of the University of Chicago. Detailed methods and analyses are available in SI Materials and Methods.
Participants.
Participants had corrected-to-normal vision and reported no history of epilepsy, brain surgery, or traumatic brain injuries. Informed consent was obtained before beginning the experiment, and participants were paid $20/h for their participation. We recorded EEG from 16 fluent signers of ASL. Data from two participants were excluded before analysis due to excessive EEG artifacts, and data from one participant were lost due to experimenter error. All participants retained in the analyses began learning ASL by 5 y of age (; 3 female, 10 male; age 24–44; mean age of acquisition 1.1 y). Participants who used hearing aids or cochlear implants removed the devices before beginning the experiment. A fluent speaker of ASL (J.L.) answered participants’ questions about the study. We recorded EEG from an additional 16 participants who had no prior exposure to ASL. These participants were recruited from the University of Chicago community through online postings. One participant who was currently learning ASL was excluded before analyses, leaving nonsigning participants (10 female, 5 male; ages 18–31).
IVC.
The IVC represents a time-series of aggregated visual change between frames (Fig. 1) and is computed as the sum of squared differences in each pixel across sequential frames of video:
where is the grayscale value of pixel at time . Python code for the IVC is available at https://github.com/gbrookshire/ivc.
EEG Analysis.
See SI Materials and Methods for details about EEG acquisition and preprocessing. To compute coherence, IVC and EEG data were filtered into overlapping log-spaced frequency bins by using phase-preserving forward-reverse Butterworth bandpass filters. Bins were centered on values from 0.5 to 16 Hz and included frequencies in the range , where is the center frequency for . Instantaneous phase and power were determined with the Hilbert transform. Power was computed as the absolute value of the analytic signal, and phase as the angle of the analytic signal. These power and phase estimates were then used to calculate coherence:
where is the time point, is the phase difference between the IVC and EEG, is power in the IVC, and is power in the EEG recording (9). Statistical significance of coherence was determined by a two-stage randomization procedure. First, the IVC was randomly shifted to obtain a null distribution of coherence between the two signals. Second, statistical significance was determined by using cluster-based permutation tests (56) (SI Materials and Methods).
SI Materials and Methods
Spectral Analysis of Speech and Sign.
We analyzed temporal structure in multiple spoken and signed languages. These analyses were performed by using custom software written in Python.
Spoken-language samples comprised audio recordings of stories in nine languages from five languages families (n = 12 samples, total duration 1:07:28; Table S1). Before computing the spectra, these recordings were trimmed to a maximum of 6 min (mean 5:37, SD 0:52), mixed down to 1 output channel, and downsampled to 22,050 Hz.
Table S1.
Spoken language samples
Sign-language samples were chosen to be as similar as possible to the spoken samples (n = 14, total duration 1:10:22, mean 5:02, SD 3:19; Table S2), representing four languages from three families. All samples comprised forward-facing views of a single speaker. In 12 videos, the speaker told a story; in the remaining 2 videos, the speaker gave an informational speech. The IVC was calculated for each video.
Table S2.
Sign language samples
Sample name | Language | Duration, s | Source | Frame rate |
Scottish Storytelling Center: BSL storytelling bursary awarded | BSL | 758 | https://www.youtube.com/watch?v=BThdtYJIdLE | 25 |
Visual BSL Workshop | BSL | 354 | https://www.youtube.com/watch?v=tBrEZ5Pz6kQ | 25 |
DIE TAUBE ZEITMASCHINE | DGS | 120 | https://www.youtube.com/watch?v=OERsJ3yq1fs | 23.976 |
Integrationsmaßnahme für taube und schwerhörige Arbeitssuchende oder arbeitslose Personen | DGS | 165 | See Open Science Framework repository(https://osf.io/5pktz/?view_only=d4a9eb7180824a5d824a83b9b7f9ef47) | 25 |
Ausbildung zur Gebärdensprachdozentin–lehrerin und zum Gebärdensprach-dozenten—lehrer | DGS | 138 | See Open Science Framework repository (https://osf.io/5pktz/?view_only=d4a9eb7180824a5d824a83b9b7f9ef47) | 50 |
Auslan Storybooks: Spider Ants | Auslan | 202 | https://www.youtube.com/watch?v=09B03W-ZqdI | 25 |
Auslan Storybooks: A Whole New Star | Auslan | 137 | https://www.youtube.com/watch?v=53FMo3vWkLk | 25 |
Auslan Storybooks: Channel 801 | Auslan | 329 | https://www.youtube.com/watch?v=95YsoZ2BOsM | 25 |
Bonnie’s Introduction | ASL | 279 | Bonnie Kraft - Tomorrow Dad Will Still Be Deaf (DVD) | 30 |
Dummy | ASL | 302 | The Rosa Lee Show (DVD) | 30 |
Mama Speaks | ASL | 250 | Bonnie Kraft - Tomorrow Dad Will Still Be Deaf (DVD) | 30 |
CODA Is My Deaf Club | ASL | 76 | Bonnie Kraft - Tomorrow Dad Will Still Be Deaf (DVD) | 30 |
Kondima | ASL | 590 | The Rosa Lee Show (DVD) | 30 |
Little Feet | ASL | 522 | The Rosa Lee Show (DVD) | 30 |
All IVC and broadband envelope recordings were resampled to a common frequency of 30 Hz, and power was normalized by dividing out the SD of each signal. Power spectra were computed by using Welch’s method. Signals were split into 2.13-s-long segments ( samples) that overlapped by 1.07 s ( samples). A Hanning window was applied to each segment, and the linear trend was removed. Fast Fourier transforms were then computed for each segment. The spectrum for each signal was obtained by averaging the spectra in all segments.
We fit sign language IVC spectra to a function using least-squares regression by transforming the equation into the linear form , with power and frequency (22). We tested for deviations from the trend using one-sample t tests.
Broadband Envelope.
The broadband envelope was computed over samples of oral speech by adapting methods from previous studies (22). First, the raw waveform was bandpass-filtered into 25 logarithmically spaced frequency bands from 0.1 to 10 kHz using least-squares filters with 501 taps. The narrowband envelope of each filtered signal was then calculated as the absolute value of the Hilbert transform. The narrowband envelopes were summed to obtain the broadband envelope for each recording.
Stimuli.
Stimuli comprised two videos of ASL storytelling: “Kondima” and “Little Feet” from the Rosa Lee Show (57). Each movie was 10 min long with a 30-Hz frame rate and depicted a native speaker of ASL telling a story against a static background. To ensure that the timing of the movie was accurately matched with EEG recordings, a small white square flashed in the corner of the display once every 30 frames of video (out of view of the participant) and was registered by a photodiode connected to the EEG amplifier. In total, the session lasted 60–90 min.
EEG Acquisition and Preprocessing.
EEG was recorded at 250 Hz by using a 128-channel net (Electrical Geodesics). Impedances were reduced to k before participants watched each story. EEG analyses were performed in Matlab by using custom software and the FieldTrip package (58). Recordings from electrodes along the face, beneath the ears, and at the base of the neck were excluded before any analysis, leaving 103 channels in all analyses. Electrode movement artifacts were manually identified and rejected by replacing the tagged region with zeros and applying a 4,000-ms half-Hann taper to each side of the artifact. Artifacts from blinks and eye-movements were identified and removed by using independent component analysis. To ensure that the IVC of each story was accurately matched to the EEG recordings, we used cubic spline interpolation to warp the IVC to the time stamps registered by the photodiode. This process simultaneously resampled the IVC from 30 to 250 Hz. Before computing coherence, EEG signals were rereferenced to the average mastoids.
EEG Statistical Testing.
Statistical significance of EEG coherence to the IVC was determined by a two-stage randomization procedure. To obtain a null distribution of coherence, the onset of the IVC was circularly shifted to a randomly selected starting point, and coherence was computed between EEG signals and the shifted IVC. This procedure preserves the spectrotemporal characteristics of both signals, but eliminates any relationship between them. For each subject, we computed 100 randomly shifted baselines. Next, we tested for significant differences between the empirical and randomly shifted coherence using a cluster-based nonparametric permutation test (56). This test looked for a difference in coherence between the empirical and randomly shifted data, contrasted with n = 10,000 permutations in which the “empirical” data were randomly selected from the group of empirical and randomly shifted traces. For each frequency and each channel, a t statistic was computed on the difference between empirical and randomly shifted data using a dependent-samples regression. The test statistic was computed as the maximum cluster size in each permutation (cluster threshold: , two-tailed). The P value was calculated as the proportion of permuted cluster statistics that were more extreme than the empirical value.
In previous studies, cortical entrainment to speech is often strongest over auditory cortex (2, 8, 11, 35) and frontal cortex (14). We tested coherence to sign language in two analogous regions of interest (ROIs): at occipital channels and at frontal channels.
To test whether knowledge of ASL influenced the strength of entrainment, we performed cluster-permutation analyses on the difference in coherence between signers and nonsigners at the occipital and frontal ROIs. For each participant, we normalized the data by computing the Z score of their empirical coherence against the randomly shifted baseline coherence values. Z-scored empirical coherence was then compared between groups using cluster permutation tests.
Data Deposition.
All EEG data and samples of speech and sign language are available on an Open Science Framework repository (https://osf.io/5pktz/?view_only=d4a9eb7180824a5d824a83b9b7f9ef47).
Supplementary Material
Acknowledgments
This work was supported by a William Orr Dingwall Neurolinguistics Fellowship (to G.B.); a research grant from the University of Chicago Center for Gesture, Sign, and Language (to G.B. and J.L.); National Science Foundation Graduate Research Fellowship Program Grant DGE-1144082 (to J.L.); the Institute of Education Sciences; U.S. Department of Education Grant R305B140048 (to J.L.); McDonnell Scholar Award 220020236 (to D.C.); National Science Foundation Grant BCS-1257101 (to D.C.); and National Science Foundation Grant BCS-0116293 (to H.C.N.). Images from The Rosa Lee Show are used courtesy of Rosa Lee Timm.
Footnotes
The authors declare no conflict of interest.
This article is a PNAS Direct Submission. E.A. is a guest editor invited by the Editorial Board.
This article contains supporting information online at www.pnas.org/lookup/suppl/doi:10.1073/pnas.1620350114/-/DCSupplemental.
References
- 1.Ahissar E, et al. Speech comprehension is correlated with temporal response patterns recorded from auditory cortex. Proc Natl Acad Sci USA. 2001;98(23):13367–13372. doi: 10.1073/pnas.201400998. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Luo H, Poeppel D. Phase patterns of neuronal responses reliably discriminate speech in human auditory cortex. Neuron. 2007;54(6):1001–1010. doi: 10.1016/j.neuron.2007.06.004. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Gross J, et al. Speech rhythms and multiplexed oscillatory sensory coding in the human brain. PLoS Biol. 2013;11(12):e1001752. doi: 10.1371/journal.pbio.1001752. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Park H, Ince RAA, Schyns PG, Thut G, Gross J. Frontal topdown signals increase coupling of auditory low-frequency oscillations to continuous speech in human listeners. Curr Biol. 2015;25(12):1649–1653. doi: 10.1016/j.cub.2015.04.049. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Schroeder CE, Lakatos P. Low-frequency neuronal oscillations as instruments of sensory selection. Trends Neurosci. 2009;32(1):9–18. doi: 10.1016/j.tins.2008.09.012. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Giraud A-L, Poeppel D. Cortical oscillations and speech processing: Emerging computational principles and operations. Nat Neurosci. 2012;15(4):511–517. doi: 10.1038/nn.3063. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Peelle JE, Davis MH. Neural oscillations carry speech rhythm through to comprehension. Front Psychol. 2012;3(320):1–17. doi: 10.3389/fpsyg.2012.00320. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Howard MF, Poeppel D. Discrimination of speech stimuli based on neuronal response phase patterns depends on acoustics but not comprehension. J Neurophysiol. 2010;104(5):2500–2511. doi: 10.1152/jn.00251.2010. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Doelling KB, Arnal LH, Ghitza O, Poeppel D. Acoustic landmarks drive delta–theta oscillations to enable speech comprehension by facilitating perceptual parsing. Neuroimage. 2014;85:761–768. doi: 10.1016/j.neuroimage.2013.06.035. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Ding N, Chatterjee M, Simon JZ. Robust cortical entrainment to the speech envelope relies on the spectro-temporal fine structure. Neuroimage. 2014;88:41–46. doi: 10.1016/j.neuroimage.2013.10.054. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Luo H, Liu Z, Poeppel D. Auditory cortex tracks both auditory and visual stimulus dynamics using low-frequency neuronal phase modulation. PLoS Biol. 2010;8(8):e1000445. doi: 10.1371/journal.pbio.1000445. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Crosse MJ, Butler JS, Lalor EC. Congruent visual speech enhances cortical entrainment to continuous auditory speech in noise-free conditions. J Neurosci. 2015;35(42):14195–14204. doi: 10.1523/JNEUROSCI.1829-15.2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Zion Golumbic EM, et al. Mechanisms underlying selective neuronal tracking of attended speech at a “cocktail party”. Neuron. 2013;77(5):980–991. doi: 10.1016/j.neuron.2012.12.037. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Kayser SJ, Ince RAA, Gross J, Kayser C. Irregular speech rate dissociates auditory cortical entrainment, evoked responses, and frontal alpha. J Neurosci. 2015;35(44):14691–14701. doi: 10.1523/JNEUROSCI.2243-15.2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Ahissar E, Ahissar M. Processing of the temporal envelope of speech. In: König R, Heil P, Budinger E, Scheich H, editors. The Auditory Cortex: A Synthesis of Human and Animal Research. Lawrence Erlbaum Associates; London: 2005. pp. 295–313. [Google Scholar]
- 16.VanRullen R, Zoefel B, Ilhan B. On the cyclic nature of perception in vision versus audition. Philos Trans R Soc Lond B Biol Sci. 2014;369(1641):20130214. doi: 10.1098/rstb.2013.0214. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Landau AN, Fries P. Attention samples stimuli rhythmically. Curr Biol. 2012;22(11):1000–1004. doi: 10.1016/j.cub.2012.03.054. [DOI] [PubMed] [Google Scholar]
- 18.Herrmann CS. Human EEG responses to 1–100 hz flicker: Resonance phenomena in visual cortex and their potential correlation to cognitive phenomena. Exp Brain Res. 2001;137(3-4):346–353. doi: 10.1007/s002210100682. [DOI] [PubMed] [Google Scholar]
- 19.Mathewson KE, et al. Making waves in the stream of consciousness: Entraining oscillations in EEG alpha and fluctuations in visual awareness with rhythmic visual stimulation. J Cogn Neurosci. 2012;24(12):2321–2333. doi: 10.1162/jocn_a_00288. [DOI] [PubMed] [Google Scholar]
- 20.VanRullen R, Macdonald JSP. Perceptual echoes at 10 Hz in the human brain. Curr Biol. 2012;22(11):995–999. doi: 10.1016/j.cub.2012.03.050. [DOI] [PubMed] [Google Scholar]
- 21.Spaak E, de Lange FP, Jensen O. Local entrainment of alpha oscillations by visual stimuli causes cyclic modulation of perception. J Neurosci. 2014;34(10):3536–3544. doi: 10.1523/JNEUROSCI.4385-13.2014. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Chandrasekaran C, Trubanova A, Stillittano S, Caplier A, Ghazanfar AA. The natural statistics of audiovisual speech. PLoS Comput Biol. 2009;5(7):e1000436. doi: 10.1371/journal.pcbi.1000436. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Meier RP. Why different, why the same? Explaining effects and non-effects of modality upon linguistic structure in sign and speech. In: Meier RP, Cormier K, Quinto-Pozos D, editors. Modality and Structure in Signed and Spoken Languages. Cambridge Univ Press; Cambridge, UK: 2002. pp. 1–26. [Google Scholar]
- 24.Petitto LA, Holowka S, Sergio LE, Levy B, Ostry DJ. Baby hands that move to the rhythm of language: Hearing babies acquiring sign languages babble silently on the hands. Cognition. 2004;93(1):43–73. doi: 10.1016/j.cognition.2003.10.007. [DOI] [PubMed] [Google Scholar]
- 25.Bellugi U, Fischer S. A comparison of sign language and spoken language. Cognition. 1972;1(2):173–200. [Google Scholar]
- 26.Hwang S-OK. 2011. Windows into sensory integration and rates in language processing: Insights from signed and spoken languages. PhD Thesis (Univ of Maryland, College Park, MD)
- 27.Greenberg S, Carvey H, Hitchcock L, Chang S. Temporal properties of spontaneous speech: A syllable-centric perspective. J Phon. 2003;31(3):465–485. [Google Scholar]
- 28.Ng BS, Schroeder T, Kayser C. A precluding but not ensuring role of entrained low-frequency oscillations for auditory perception. J Neurosci. 2012;32(35):12268–12276. doi: 10.1523/JNEUROSCI.1877-12.2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Henry MJ, Obleser J. Frequency modulation entrains slow neural oscillations and optimizes human listening behavior. Proc Natl Acad Sci USA. 2012;109(49):20095–20100. doi: 10.1073/pnas.1213390109. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Hickok G, Farahbod H, Saberi K. The rhythm of perception: Entrainment to acoustic rhythms induces subsequent perceptual oscillation. Psychol Sci. 2015;26:1006–1013. doi: 10.1177/0956797615576533. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Hanslmayr S, et al. Prestimulus oscillations predict visual perception performance between and within subjects. Neuroimage. 2007;37(4):1465–1473. doi: 10.1016/j.neuroimage.2007.07.011. [DOI] [PubMed] [Google Scholar]
- 32.van Dijk H, Schoffelen J-M, Oostenveld R, Jensen O. Prestimulus oscillatory activity in the alpha band predicts visual discrimination ability. J Neurosci. 2008;28(8):1816–1823. doi: 10.1523/JNEUROSCI.1853-07.2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Mathewson KE, Gratton G, Fabiani M, Beck DM, Ro T. To see or not to see: Prestimulus phase predicts visual awareness. J Neurosci. 2009;29(9):2725–2732. doi: 10.1523/JNEUROSCI.3963-08.2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Busch NA, Dubois J, VanRullen R. The phase of ongoing EEG oscillations predicts visual perception. J Neurosci. 2009;29(24):7869–7876. doi: 10.1523/JNEUROSCI.0113-09.2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Park H, Kayser C, Thut G, Gross J. Lip movements entrain the observers’ low-frequency brain oscillations to facilitate speech intelligibility. eLife. 2016;5:e14521. doi: 10.7554/eLife.14521. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Power AJ, Mead N, Barnes L, Goswami U. Neural entrainment to rhythmically presented auditory, visual, and audio-visual speech in children. Front Psychol. 2012;3:216. doi: 10.3389/fpsyg.2012.00216. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Cravo AM, Rohenkohl G, Wyart V, Nobre AC. Temporal expectation enhances contrast sensitivity by phase entrainment of low-frequency oscillations in visual cortex. J Neurosci. 2013;33(9):4002–4010. doi: 10.1523/JNEUROSCI.4675-12.2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Ilhan B, VanRullen R. No counterpart of visual perceptual echoes in the auditory system. PLoS One. 2012;7(11):e49287. doi: 10.1371/journal.pone.0049287. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.O’Sullivan JA, et al. Attentional selection in a cocktail party environment can be decoded from single-trial EEG. Cereb Cortex. 2015;25(7):1697–1706. doi: 10.1093/cercor/bht355. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Ghazanfar AA, Takahashi DY, Mathur N, Fitch WT. Cineradiography of monkey lip-smacking reveals putative precursors of speech dynamics. Curr Biol. 2012;22(13):1176–1182. doi: 10.1016/j.cub.2012.04.055. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Ghazanfar AA, Morrill RJ, Kayser C. Monkeys are perceptually tuned to facial expressions that exhibit a theta-like speech rhythm. Proc Natl Acad Sci USA. 2013;110(5):1959–1963. doi: 10.1073/pnas.1214956110. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Doelling KB, Poeppel D. Cortical entrainment to music and its modulation by expertise. Proc Natl Acad Sci USA. 2015;112(45):E6233–E6242. doi: 10.1073/pnas.1508431112. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Lakatos P, Karmos G, Mehta AD, Ulbert I, Schroeder CE. Entrainment of neuronal oscillations as a mechanism of attentional selection. Science. 2008;320(5872):110–113. doi: 10.1126/science.1154735. [DOI] [PubMed] [Google Scholar]
- 44.MacSweeney M, Capek CM, Campbell R, Woll B. The signing brain: The neurobiology of sign language. Trends Cogn Sci. 2008;12(11):432–440. doi: 10.1016/j.tics.2008.07.010. [DOI] [PubMed] [Google Scholar]
- 45.Leonard MK, et al. Signed words in the congenitally deaf evoke typical late lexicosemantic responses with no early visual responses in left superior temporal cortex. J Neurosci. 2012;32(28):9700–9705. doi: 10.1523/JNEUROSCI.1002-12.2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Corina DP, Jose-Robertson LS, Guillemin A, High J, Braun AR. Language lateralization in a bimanual language. J Cogn Neurosci. 2003;15(5):718–730. doi: 10.1162/089892903322307438. [DOI] [PubMed] [Google Scholar]
- 47.Gutierrez-Sigut E, et al. Language lateralization of hearing native signers: A functional transcranial doppler sonography (ftcd) study of speech and sign production. Brain Lang. 2015;151:23–34. doi: 10.1016/j.bandl.2015.10.006. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48.Hickok G, Bellugi U, Klima ES. The neural organization of language: Evidence from sign language aphasia. Trends Cogn Sci. 1998;2(4):129–136. doi: 10.1016/s1364-6613(98)01154-1. [DOI] [PubMed] [Google Scholar]
- 49.MacSweeney M, Waters D, Brammer MJ, Woll B, Goswami U. Phonological processing in deaf signers and the impact of age of first language acquisition. Neuroimage. 2008;40(3):1369–1379. doi: 10.1016/j.neuroimage.2007.12.047. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.Williams JT, Darcy I, Newman SD. Modality-independent neural mechanisms for novel phonetic processing. Brain Res. 2015;1620:107–115. doi: 10.1016/j.brainres.2015.05.014. [DOI] [PubMed] [Google Scholar]
- 51.Newman AJ, Supalla T, Hauser P, Newport EL, Bavelier D. Dissociating neural subsystems for grammar by contrasting word order and inflection. Proc Natl Acad Sci USA. 2010;107(16):7539–7544. doi: 10.1073/pnas.1003174107. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 52.Emmorey K, Mehta S, Grabowski TJ. The neural correlates of sign versus word production. Neuroimage. 2007;36(1):202–208. doi: 10.1016/j.neuroimage.2007.02.040. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 53.Emmorey K, McCullough S, Mehta S, Grabowski TJ. How sensory-motor systems impact the neural organization for language: Direct contrasts between spoken and signed language. Front Psychol. 2014;5:484. doi: 10.3389/fpsyg.2014.00484. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 54.Jacobs J, Kahana MJ, Ekstrom AD, Fried I. Brain oscillations control timing of single-neuron activity in humans. J Neurosci. 2007;27(14):3839–3844. doi: 10.1523/JNEUROSCI.4636-06.2007. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 55.Lakatos P, et al. An oscillatory hierarchy controlling neuronal excitability and stimulus processing in the auditory cortex. J Neurophysiol. 2005;94(3):1904–1911. doi: 10.1152/jn.00263.2005. [DOI] [PubMed] [Google Scholar]
- 56.Maris E, Oostenveld R. Nonparametric statistical testing of EEG-and MEG-data. J Neurosci Meth. 2007;164(1):177–190. doi: 10.1016/j.jneumeth.2007.03.024. [DOI] [PubMed] [Google Scholar]
- 57.Timm RL, Timm D. 2008. The Rosa Lee Show: 2004–2008 [DVD] (Pelican Ave Inc., Fremont, CA). American sign language performance art.
- 58.Oostenveld R, Fries P, Maris E, Schoffelen J-M. FieldTrip: Open source software for advanced analysis of MEG, EEG, and invasive electrophysiological data. Comput Intell Neurosci. 2010;2011:156869. doi: 10.1155/2011/156869. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.