Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2021 Jul 1.
Published in final edited form as: Int J Lang Commun Disord. 2020 Apr 29;55(4):445–457. doi: 10.1111/1460-6984.12535

Understanding Event-Related Potentials (ERPs) in Clinical and Basic Language and Communication Disorders Research: A Tutorial

Sean McWeeny 1, Elizabeth S Norton 1,2
PMCID: PMC7802513  NIHMSID: NIHMS1658739  PMID: 32347637

Abstract

Background

Event-related potentials (ERPs), which are electrophysiological neural responses time-locked to a stimulus, have become an increasingly common tool in language and communication disorders research. ERPs can provide complementary evidence to behavioral measures as well as unique perspectives on communication disorders. ERPs have distinct advantages of providing precise information about timing of neural processes and can be used in cases where it is difficult to obtain responses from participants, such as infants or individuals who are minimally verbal. However, clinicians and clinician-scientists rarely receive training in how to interpret ERP research.

Aims

The aim of this review is to provide information that allows readers to better understand, interpret, and evaluate research using ERPs. We focus on research related to communication sciences and disorders and the information that is most relevant to interpreting research articles.

Method

We explain what ERPs are and how ERP data are collected, referencing key texts and primary research articles. Potential threats to validity, guidelines for interpreting data, and the pros and cons using of ERPs are discussed. Research in the area of pediatric language disorders is used as a model; common paradigms such as the semantic incongruity N400 and the auditory mismatch negativity are used as tangible examples. With this foundation of understanding ERPs, the state of the field in terms of how ERPs are used and the ways that they may inform the field are discussed.

Main Contribution

To date, no review article has focused on ERPs as they relate to clinical or communication research. The main contribution of this article is that it provides practical information geared toward understanding ERP research.

Conclusions

ERPs offer insights into neural processes supporting communication and can both complement behavior and provide information that behavioral measures cannot. We encourage readers to critically evaluate articles using ERPs, effectively pushing the field forward through increased understanding and rigor.

Keywords: event-related potential, ERP, neuroscience, MMN, N400

Introduction

Communication requires a host of complex neural processes, many of which are still not fully understood. When the brain is processing a stimulus, like a spoken or written word, it produces characteristic electrical signals that can be recorded from sensors on the scalp. These neural signals in response to stimuli, called event-related potentials (ERPs), can index various sensory, cognitive, and motor processes such as detecting a stimulus, evaluating or categorizing it, detecting an error or anomaly in the stimulus, and making a motor response such as pressing a button or articulating a word. ERPs are essentially measurements of small electrical voltage changes produced by the brain, with very precise timing. ERPs can provide insights that behavioral measures cannot, such as what happens just milliseconds after a stimulus is presented (before a behavioral response can occur), or processing that happens subconsciously (without explicit effort or attention). They can also provide information into the etiology of disorders; for example, if a group of patients with a certain disorder look like typical controls in terms of their ERPs that reflect early sensory processing but different in their later appraisal of stimuli, that indicates that the disorder may cause disruption after the phase of early sensory processing. ERPs are also very useful for studying populations who may not be able to make overt responses or follow directions in behavioral tasks, such as infants and young children or patients who have motor, language, or cognitive deficits. Assessing the brain can also provide information on the etiology of communication disorders and may reveal biomarkers of a disorder that behavioral measures alone cannot. For these and other reasons, ERP studies have become increasingly common in speech, language, and hearing research.

ERPs can be challenging to understand and interpret without some background in this area, which is not typically part of clinical education and training in communication sciences and disorders. The primary goal of this tutorial is thus to provide speech, language, and hearing clinicians and scientists with a foundation from which they can understand and evaluate research using ERPs. As with any method, ERP research needs to be rigorously assessed not only by the peer-review process, but also by the “consumers” of scientific research. In the same way that a reader needs to understand an intervention approach or statistical procedure to fully evaluate a research paper that uses it, understanding the basics of ERPs can help readers evaluate the conclusions of papers that use them. The goals of this tutorial are to describe: (1) what ERPs are and where they come from, (2) suggestions to evaluate the quality and validity of an ERP publication’s methods and interpretation of data, and (3) how ERPs can be effectively used in research with clinical relevance, using work in the area of pediatric language as an example.

Note that there are several methods that are similar or related to ERPs that are outside of the scope of this review. Electroencephalography (EEG) is the continuous signal from which ERPs are derived, and often involves measuring the oscillatory activity of the brain (for a review of EEG in cognitive development see Bell and Cuevas 2012, for review of EEG and language, see Weiss and Mueller 2003). Magnetoencephalography (MEG) is a measure of magnetic changes in the brain, similar to EEG (for a review of MEG/EEG methods, see Puce and Hämäläinen 2017). Electrocorticography (ECoG) records electrical activity directly from the brain’s surface (for a review see He 2015). Also outside the current scope are the auditory brainstem response (ABR) and frequency following response (FFR); although both can be considered ERPs, their interpretation and characteristics differ sufficiently to exclude from this tutorial (for a review see Skoe and Kraus 2010). Because the ABR’s neural origins are known, and it is highly reliable/replicable, it has been extremely clinically useful; building on years of research, newborn hearing screening using ABR is common and required across most of the US (National Center for Hearing Assessment and Management 2011).

What is an Event-Related Potential (ERP)?

Recall that an ERP is a measurement of the brain’s electrical response to a stimulus that is recorded at the scalp. However, to understand how to collect those measurements, we must first consider how electrical activity is recorded from the living human brain. Electrodes (electrical sensors) are placed on the scalp to record EEG, which is a continuous measure of the electrical activity of the brain, similar to an EKG of the heart. ERPs represent average activity from the continuous EEG signal that is synchronized (i.e., time-locked) to a certain stimulus (e.g., an auditory tone or a printed word).

The electrical activity recorded with EEG comes from neurons in the brain, which function by creating electrical signals. The electrical signals from a neuron are tiny in magnitude, so the signals that can be measured scalp reflect thousands of neurons changing their electrical activity in similar patterns. ERPs mostly reflect a certain type of activity from one type of neuron arranged in a certain layout in the cortex, (postsynaptic potentials of cortical pyramidal neurons oriented perpendicular to the scalp; for more details on the neural origin of ERPs, see Buzsáki et al. 2012, Luck 2014) but understanding this is not paramount to understanding most applications of ERPs.

The amplitude of electrical activity recorded over time is displayed as brainwaves (typically with a measure of voltage on the Y axis and time on the X axis) and serves as the basis for EEG and ERP signals. To get from continuous EEG to the brain’s electrical response to an event (i.e., an ERP), the EEG signal must be time-locked to a stimulus. However, the signal from any one stimulus, or event, captures brain signals that resulted from that event, but also a substantial amount of noise (this includes unrelated neural processes, measurement error, and environmental electrical noise). Even in perfect recording conditions with no noise from measurement error or the environment, unrelated brain signals are captured in the recording. Thus, presenting the stimulus many times and averaging the responses together isolates the activity that is related to the event of interest.

ERP Components and Paradigms

What Are ERP Components?

Experimental paradigms for ERPs use stimuli that elicit a series of components, which are sensory, perceptual, cognitive, or motor processes that appear as a series of positive and negative changes in the waveform (the plotted ERP voltage) over time. Put simply, a specific ERP component is the brain’s electrical response to a specific stimulus feature. Most components are named for the direction of their voltage deflection (P for positive, N for negative) and the approximate timing (e.g., N400 component peaks around 400ms) or order of the peak (N1 for first negative-going peak), though others are abbreviations of their names (such as the MMN, or mismatch negativity). Components have an onset, a peak or plateau, and an offset. For example, the onset of a component might be when the first few thousand neurons in primary auditory cortex are systematically responding to a sound, and the offset would be when auditory cortex quiets to baseline levels.

ERP components reflect how processing in the brain unfolds over time, from early sensory processing components precede later cognitive processing components that require attention and evaluation of the stimulus. Early components reflecting sensory and perceptual processing of a stimulus are often called “pre-attentive,” meaning that they occur without or before conscious attention from the participant. Later components typically reflect cognitive processing and require conscious attention, though some cognitive ERPs can appear without the subject consciously perceiving a stimulus (Luck et al. 1996). Passive or unattended paradigms are well-suited to pediatric populations, as they do not require any overt responses or compliance with directions, and they can be collected while the child is attending to something more engaging, like watching a movie. However, well-designed and engaging paradigms that require active participation or attention have been successfully utilized in numerous ERP experiments with children (e.g., Henderson et al. 2011).

Examples of ERP Paradigms to Guide the Tutorial

Throughout this tutorial, we will use two classic ERP paradigms in the area of language processing to give concrete examples of concepts. We will first introduce these experimental paradigms and then how the ERP that they elicit is measured.

Semantic incongruity paradigm and N400 component.

The N400 semantic incongruity paradigm based on a landmark paper in which Marta Kutas and Steve Hillyard established that ERPs can index detection of a semantic incongruity in language (Kutas and Hillyard 1980). In this now widely-used paradigm, a participant reads a sentence that is presented one word at a time on a computer screen; most of the sentences are semantically appropriate or expected, but some have an incongruous or unexpected final word. An example semantically appropriate sentence would be “It was his first day at work.” In the original experiment, 25% of the sentences had a semantic incongruity, that is, the final word of the sentence is unexpected from the context, such as “He spread the warm bread with socks.” The brain’s response to the final word is very different in this semantically incongruous condition than in the expected condition. This ERP response to semantic incongruity is called the N400 component, as there is a more negative voltage for the unexpected word around 400ms after the word is presented. The N400 component also occurs when a word is semantically acceptable but not expected. For example, the sentence “Do not touch the wet dog.” would elicit an N400, even though it is not a semantic error; “Do not touch the wet paint.” is the semantically expected ending (Kutas and Hillyard 1984). The N400 occurs regardless of whether the stimuli were presented visually (e.g., Kutas and Hillyard 1980) or auditorily (e.g., McCallum et al. 1984) and thus specifically relates to lexical-semantic processing. The N400 may be used to answer questions about the strength of semantic relationships, or whether clinical populations have different patterns of semantic processing. For example, studies have used the N400 component in a modified semantic paradigm to examine whether receptive semantic processing differs in children with autism spectrum disorder who are minimally verbal or nonverbal (e.g., Cantiani et al. 2016).

Auditory oddball paradigm and mismatch negativity (MMN) component.

The second example is an oddball paradigm in which one stimulus is presented often and one or more other stimuli are presented rarely. For example, the syllable /ba/ is played 90% of the time and the syllable /da/ is played the remaining 10% of the time while the participant watches a silent video. The difference between the frequent stimulus (/ba/) and the oddball or deviant stimulus (e.g., /da/) is a component known as the mismatch negativity, or MMN (Näätänen and Kreegipuu 2011). The rare sound elicits a larger negative voltage response from about 100–250ms; this difference between the rare and frequent stimuli, shown in figure 1A, is the MMN. If the stimuli probabilities were switched, (e.g., /ba/ played 10% and /da/ played 90%), we would see a highly similar pattern of rare vs. frequent responses as in the original paradigm, indicating that the differences aren’t solely driven due to acoustic differences in the two syllables but depend on more general change detection processes. The MMN can be used to index automatic auditory discrimination of different auditory features (such as tone pitch or duration) and how the brain tracks stimulus probability, which are skills important for language, especially language learning. If the difference between the two sounds is too small for the participant to perceive, the MMN is not observed.

Figure 1. Typical plots present in ERP research.

Figure 1.

A) Group grand average waveforms (averaged from about 100 child participants) showing 2 conditions (frequent and rare syllables, in black and red) and the difference or subtraction between them (in blue). B) Group grand average scalp plot showing spatial distribution of the mean amplitude of the difference wave depicted in (A) within a time window of interest from 300–500ms. The negativity shown in blue extends bilaterally across fronto-central electrode sites. (Unpublished data from Author’s Lab.)

This is only one example of an oddball paradigm, which is a commonly used ERP paradigm that can be modified with multiple deviants, different stimulus probabilities (e.g., 25% deviants), or a task where the participant actively categorizes the stimuli. The MMN allows researchers to investigate populations in which auditory discrimination may be impaired. Because it does not require overt attention or responses, it has been widely used in infants and children (Bishop 2007) and shown to differ in groups with or at risk for language and reading disorders (Bishop 2007, Bruder et al. 2011). Although it is most commonly used in auditory research, some evidence exists that it exists for visual stimuli as well (Kimura et al. 2009).

Other Common ERP Components

Although it is not necessary to have in-depth knowledge of established components to read ERP research, understanding these components will give you a better sense of how they may be applied to research. These descriptions are designed to give a cursory overview of what types of stimuli and paradigms may elicit components related to communication. There are many more components that not covered here, such as ones related to syntactic incongruity (P600) or error detection (the error-related negativity, ERN). Note that some auditory and visual components share the same name, despite originating from completely different areas of the brain and responding to different stimulus modalities (i.e., hearing vs. vision). For further detailed reading, see the Oxford Handbook of ERP Components (Luck and Kappenman 2011).

Auditory sensory components.

In adults, cortical auditory responses begin with a positive wave around 50ms called the P1 or P50 component. The P50 is generated in auditory cortex and reflects sensory gating (Pratt 2011), the process through which the brain filters out irrelevant information. P50 is quickly followed by a larger negative wave called the N1, or N100. The N100 reflects a change in auditory stimulation, like the offset and onset of sounds (Pratt 2011). Some evidence suggests that N1 may be used to index speech segmentation, such as in a statistical learning paradigm (Sanders et al. 2002). In young children these components are different, with P1 being larger in children than adults (Čeponiené et al. 2002); across development, they tend to shift in time and change in shape.

Visual sensory components.

Although brief flashes of visual stimuli can evoke ERPs as soon as 70ms after stimulus onset (Allison et al. 1977), the main visual ERPs described in research manifest as a series of three peaks. The first of these components, C1, changes polarity based on where in the visual field the stimulus is presented (hence its name having a C rather than a P or N). It is followed by P1 and N1, which overlap in time (in other words, P1 hasn’t completely offset by the time N1 begins); this a similar pattern to auditory P1 and N1, though from different neural sources. Although C1, N1, and P1 components each peak at different times, they are all generated in visual cortex in short succession (Pratt 2011). All are affected by luminance (i.e., brightness) change, like a white stimulus appearing on a black background. These three components nicely demonstrate that components can have temporal overlap in their offsets and onsets and thus potentially influence the peak latency of any of the three components (demonstrated in figure 2).

Figure 2. How different underlying component structures can manifest as identical waveforms.

Figure 2.

A) A theoretical ERP waveform with 3 peaks. B) One possibility for the underlying component structure that when summed together, creates the waveform in panel A. C) Another possibility for the component structure of the waveform in panel A. Notice the shorter duration of X2 as compared to X’2 and the larger amplitude of X3 as compared to X’3 in panel B.

P300/P3/P3a/P3b.

Like the MMN, the P3 (also called P300) is typically elicited by an oddball paradigm. However, unlike the MMN, it requires the participant to pay attention to all the stimuli and have an expectation or categorization of the difference between the standard and the rare target stimuli (Polich 2007). It can be elicited using stimuli from visual, auditory, and other sensory domains (e.g., tactile, Brouwer and Van Erp 2010, Polich 2007). The P3 wave can be broken up into two unique subcomponents: P3a and P3b. These two subcomponents reflect different aspects of the stimulus; P3b occurs only when the target stimulus is presented (the one the participant is trying to detect), but P3a appears to an infrequent stimulus, regardless of whether the participant was looking for it or not (Polich 2011). The exact cognitive processes that P3b represents (e.g., working memory vs. a more general process) remains an active area of research and debate (Polich 2011). This case illustrates how a single large wave may have multiple generators or underlying subcomponents. The P3 component may be useful in cases in which participants are classifying objects, for example, based on certain object features (Deveney et al. 2019).

ERPs from Recording to Analysis

A familiarity with the basics of the experimental set-up can help contextualize ERP data. Here, we explain how ERP data are recorded and introduce key terms that will show up in most ERP research.

Recording ERPs

To record ERPs, a cap or net holding the electrodes is placed on the scalp. Saline solution or gel is typically applied to conduct the electrical signal from the scalp to the electrodes. This typically takes 5 to 20 minutes depending on the system and participant. (Newer “dry” electrode systems that do not require conductive gel or liquid are becoming available, but it is not yet clear whether these systems can provide signal-to-noise ratio that is on par with existing research-quality EEG systems.) Net and cap systems measure the same signals but have some advantages and trade-offs. Net systems can be quicker to set up as the saline solution can be applied to all electrodes at once and then just placed on the participant’s head. Electrodes in a cap system need to be gelled individually which takes more time, but the data are often cleaner (less noisy), especially with active electrode systems that have signal amplifiers at each site. In the past, the scalp was typically abraded (scraped) to remove a layer of skin to improve signal, but this is not typical with modern systems, which makes EEG recording safer and more comfortable for participants. Additional “external” electrodes are typically placed near the eyes to detect eye movements and blinks, as the associated muscle movements generate an electrical signal which can contaminate the data of other nearby electrodes. Very young children may not tolerate electrodes on their face, and so experimenters may use electrodes from the forehead to detect blinks. We discuss how to clean blinks (i.e., reducing their effect on the data) in the Preparing Data for Analysis section. External electrodes may also be placed to record a reference or comparison signal, such as on the earlobe or mastoid bones behind the ears. All electrodes are typically plugged in to a small box that amplifies and/or converts the signal to a format that can be recorded by a computer. Unlike MRI or MEG, this equipment is small and can be portable.

Electrodes are typically named and labeled based on their location using what is called the 10–20 system, which provides a standardized map for locating an electrode on the head across people and studies (Homan et al. 1987). Electrode locations in the 10–20 system are labeled with a letter indicating where they are on the head (e.g., Fp, frontal pole; F, frontal; FC, fronto-central; C, central; CP, centro-parietal; P, parietal; PO, parieto-occipital; O, occipital; T, temporal; FT, fronto-temporal; TP, temporo-parietal) and number (even numbers on the right, odd on the left, with larger numbers indicating farther distance from the midline, which uses z instead of a number). For example, Fz is located over frontal cortex, equidistant from the ears, and F1 is slightly to the right of Fz over the right hemisphere.

Once the electrodes are placed and connected, the experiment can begin. Researchers typically use a software program running on a computer to present stimuli to the participant. Every time a visual or auditory stimulus is presented (though ERPs exist for other sensory modalities, they are rare), a code is sent from the presentation software to the EEG acquisition software that is recording data from the electrodes. This code in the EEG data indicates the exact moment a stimulus was presented, and different codes are typically used to indicate the type of stimulus that was presented. Using the semantic incongruity task example, if we know exactly when each semantically appropriate word and when each semantic incongruity was presented, we can isolate the electrical activity to each type of stimulus (i.e., incongruous or not) by averaging the trials of each type together.

Time and Voltage: The Units of ERP Measurements

ERPs reflect how electrical voltage changes over time, typically over several hundred milliseconds. Voltage and time also characterize the two primary measurements of ERPs: amplitude and latency. Amplitude is how large an ERP voltage is, and latency is when an ERP happens; we discuss how to best measure these later. ERPs measure voltage over time with incredible temporal resolution, often hundreds or even thousands of times per second. This temporal precision is perhaps the greatest strength of ERPs. Because electrical current travels very quickly and because the brain processes certain aspects of a stimulus very quickly, the human auditory brainstem can register sound about 1.5 milliseconds after it is played. A cognitive event like detection of a semantic incongruity occurs much later 400ms post-stimulus onset, though that may not seem long!

Most research labs use systems that record voltage from many electrodes simultaneously; 32 and 64 electrode caps are common in research, and nets often include 128 or 256 electrodes. (More electrodes are not necessarily always better; see Luck, 2014 Chapter 5). Voltage at each electrode is compared to activity at another site such as the nose, earlobe, mastoid processes, or an average of all the scalp electrodes. This comparison process is known as referencing, and the site of the reference electrode(s) greatly affects how the data appear. For example, using Cz (the electrode at the very top of the head) as the reference would produce different looking ERPs than the average of all electrodes as a reference, as the reference is subtracted from each electrode. Studies with similar paradigms typically will use the same reference locations as previous work, meaning that you can compare directly across publications; however, when comparing results across publications, check to see if studies use the same reference electrodes.

Preparing Data for Analysis

ERP analysis involves many steps and many choices, and so understanding the main aspects of analysis is important for interpreting results from papers. There are established best practices for making these decisions and guidelines for what information about ERP processing that studies need to report (Keil et al. 2014). Studies that do not follow these guidelines could report insufficient information for replication, inaccurate or biased data, or inaccurate conclusions about neural differences. Thus, this section aims give readers some tools to better understand the steps and decisions that go in to making inferences about ERPs and the relative methodological strength of a given publication.

There are several steps needed to go from an EEG recording to a set of ERP waveforms. Here, we describe a typical processing “pipeline” for generating ERP measurements from EEG data. The first of these steps is to filter the EEG data so that only activity that is likely to be relevant to brain processes remains. Much of the activity of neurons is rhythmic, or oscillatory. Thus, a first step is to filter out oscillatory activity that is slower than about .01–.3 Hz (high pass filter), or faster than about 100 Hz (low pass filter). High-pass filters at .3 Hz and above can systematically change the shape and timing of ERPs (Tanner et al. 2015), so that is something for readers to watch for. Often, a notch filter, which knocks out a single frequency or small frequency range, is applied to remove electrical noise (e.g., at 50 Hz in Europe and Asia or 60 Hz in North America), as utility electricity is carried at those frequencies and electrodes may pick up that frequency. Electrical noise, also called line noise, is a type of artifact.

An artifact is best described as a non-event-related signal, or noise, which can arise for many reasons. To obtain the cleanest ERPs, artifacts must first be detected, and then either corrected or rejected from the data. Perhaps the most common artifact comes from the participant blinking, which creates a large voltage change due to the muscle activity required to blink. Blinks can be corrected (that is, their effect subtracted from the data) through a computational process called independent component analysis (ICA). An alternative approach is to just exclude (“reject”) moments of data with blinks; however, ICA can increase the number of usable trials, which is typically desirable.

Once ICA is complete, the data are segmented into epochs, which are periods around an event or stimulus onset. Epochs are created based on the codes that are sent when a stimulus is presented, described in the Recording ERPs section. Typically, within an experiment, each epoch is the same length. Epochs typically begin 50–200 ms before the stimulus and last until some time before the next stimulus is presented. The time before each stimulus is called the pre-stimulus baseline, or just baseline. The average voltage over each trial’s baseline is subtracted from the post-stimulus signal in an effort to reduce noise.

Next, any remaining artifacts must be detected in the data and rejected. Artifact rejection and correction is necessary, as large artifacts will distort the final ERP waveforms with large voltage changes due to noise. (The saying “a few bad apples spoil the bunch” applies in this case.) Many artifacts in EEG data come from motion by the participant; this is particularly true for children, who tend to move more during recording. Another common type of artifact appears when an electrode’s connection to the scalp was poor or lost during the experiment. The information from these electrodes can be deleted and re-created using an average of the electrodes surrounding it in a process known as interpolation. Ideally, this occurs for very few electrodes, as interpolating too many adjacent electrodes will result in an interpolated channel that is likely very different than the actual brain activity that was occurring at that electrode site. After any bad electrodes are interpolated, the data are examined for remaining artifacts. There are lots of ways to do this; one is to measure how much the measured voltage changes over a 200ms interval and reject any data that exceeds a given voltage threshold (e.g., 150 μV, or microvolts). This process is then repeated for every 200ms interval in each epoch. Some researchers choose apply the same threshold for artifact detection to each subject’s data, which can speed up this process; others choose to use individualized criteria, as “clean data” looks different in each person, which is an acceptable and often necessary approach for pediatric data (Luck 2014). The data marked as artifacts are then excluded from further analysis. It is common in studies of children for up to 50% of epochs to have artifacts. Thus, it is important to consider how many usable trials are averaged together; too few trials may lead to noisy averages that don’t reflect the real neural activity accurately. If different conditions or groups are being compared, it is ideal to make sure that conditions or groups do not have disparate numbers of included trials (see Number of trials section).

After the data have been cleaned of artifacts, epochs within each condition (e.g., all incongruous trials) are averaged together. As discussed earlier, averaging together multiple epochs allows us to see the components of interest, because random noise not associated with processing the stimulus will cancel out. Once epochs of a given type are averaged together, the resulting waveform is considered an event-related potential. Average ERPs of a whole group of participants or comparisons of ERPs between groups are often the main interest in ERP research; thus, publications don’t often include the individual participant waveforms calculated in the previous step. Instead, they typically present grand average waveforms which are calculated by averaging the waveforms from every individual in a group. For example, if we were interested in whether early auditory processing differs for toddlers at risk for dyslexia compared to those not at risk, we might compare the grand average for the at-risk group to the grand average for the not-at-risk group (e.g., Plakas et al. 2013). These smooth grand average waveforms are not necessarily what any individual’s waveforms look like, nor what any given trial might look like for any individual. This is because it takes many trials to cancel out unrelated activity (i.e., noise), and many participants to sufficiently cancel out noise and smooth the group’s response (i.e., signal). Grand averages can also be compared across different experimental conditions within-subjects, such as semantically expected vs. incongruous words, to better understand basic neural processes of semantic understanding.

Plotting ERPs and Interpreting Figures

Plotting ERPs allows us to better visualize the differences between conditions (e.g., semantic incongruity vs. semantically appropriate word) or between groups (e.g., typical development vs. autism). Typically, just before plotting, a 30-Hz low-pass filter is applied to remove high-frequency noise, which makes the waveform appear smoother. Figure 1A shows a typical ERP waveform plot using the auditory oddball paradigm (/ba/ presented 90% of the time and /da/ presented 10% of the time). Time in milliseconds is plotted on the x-axis, and voltage in microvolts in plotted on the y-axis. Here, negative values are plotted up; some figures use this convention, which has been typical in the history of ERP research, although others plot positive up, which is more intuitive for most readers. It is important to pay attention to the axis labels to confirm which direction is plotted up. The x-axis in this figure spans from −100 to 500ms relative to the stimulus onset. As discussed earlier, the average voltage over the baseline from −100 to 0ms has been subtracted from the whole waveform, and the baseline section is thus relatively flat. Readers can check for pre-stimulus baselines to be relatively flat compared to the waveforms that occur after stimulus onset; if they are not, this could indicate noisy data or inaccurate processing.

In the figure, there are three ERP waveforms plotted, as indicated in the figure legend; the waveform to the /ba/ (90%) condition is in black and the ERP waveform to the /da/ (10%) condition is in red. Notice that both the black and red waves have a positive-going wave around 100ms. Before stimulus onset and early in time, the two conditions do not differ, which suggests that neural responses are similar between conditions, as we would expect. From 100–200ms and from 250–500ms there are negative-going peaks in this wave, indicating that one condition had a stronger negative response than the other. The blue waveform is a difference wave; is a subtraction between the two conditions (deviant minus standard waves), which allows readers to more easily visualize the difference between the conditions. The waveforms being compared in a difference wave should have an approximately equivalent number of trials (Luck 2005). From our oddball paradigm example, comparing every frequent sound with every rare (deviant) sound would result in signal that is much noisier for the rare sound, because there are 9 times fewer trials of this kind presented. In this case, researchers can use a subset of the frequent sounds (for example, average together and analyze only some standards).

ERPs can also be plotted using scalp maps, which show how positive or negative the ERP signal is across the scalp at a certain time or time interval. In figure 1B, the mean amplitude (voltage) of the difference wave between 300–500ms is plotted at each electrode and then smoothed in space to create a continuous map of voltage. The color bar indicates that negative values are plotted in blue; meaning that over the frontal and central part of the scalp, the difference wave is negative. Even though we see an ERP over a certain part of the scalp, this doesn’t necessarily mean that the process happened in that brain area; this will be discussed in depth in the Localization section. Nonetheless, differences in scalp maps between conditions or groups can provide meaningful information.

Quantification or Measurement of ERPs

Although looking at the waveforms provides a great deal of information, quantifying or measuring an ERP component is necessary to statistically compare between groups or conditions. Understanding the general strengths and weaknesses of different measurement approaches helps readers to better understand the relative strength of a given paper. As mentioned earlier, ERPs can be quantified in terms of amplitude and latency; most papers measure one or both of these characteristics of a given component. Though many studies measure features about the peak of a component, there are some drawbacks to doing so. There is nothing inherently meaningful about the time and exact voltage at which a component reaches a maximum voltage (Luck 2014). Furthermore, the exact time and voltage at which peaks occur (i.e., peak latency and peak amplitude) can change based on how much noise is in the EEG signal. This is because noise distorts the waveform, which may create a false peak that does not accurately represent the component. Although peak amplitude and peak latency may provide information about a component, some less-biased alternatives are mean amplitude between two time points. Finding an alternative to peak latency that is robust to noise is more challenging; the approach that is most valid (from options such as fractional area latency, or relative criterion technique) may vary depending on the component of interest (Kiesel et al. 2008). Overall, peak measures should be regarded with these considerations in mind.

Even for noise-robust measurements such as mean amplitude, the measurement window and electrode(s) to measure and analyze must be chosen carefully. Ideally, all publications would use measurement windows selected in advance of analysis based on closely related previous literature so as to be unbiased; however, literature can differ in exactly what time window they quantified their components, or a paradigm or population may have never been studied before. When certain ERP components begin can depend on the duration of a stimulus, the time between stimulus onsets (i.e., stimulus onset asynchrony), as well as the time between stimulus offset and the next stimulus onset (i.e., inter-stimulus interval). Thus, the time when the component occurs will vary from study to study depending on the stimuli used.

Given the concerns about rigor and reproducibility in neuroscience (e.g., Loken and Gelman 2017), a major concern for all ERP research is the potential for experimenter bias, either conscious or unconscious. Ideally, analytic decisions such as which electrode site and time window to test, and how to exclude artifacts and outliers, are made in advance of data collection, so that the data cannot influence results. Practices that should be avoided are “fishing” for a time window in which there is a difference and then conducting statistical tests only in that window, or testing many time windows and electrodes and not reporting these tests and appropriately controlling for the multiple statistical comparisons. Luck and Gaspelin (2017) illustrate this nicely with an example where statistically significant results were obtained from real ERP data by pure chance by using analytic practices that are common in the field but that are not rigorous. Their message is clear: in highly multidimensional ERP data, it is extremely easy to find statistically significant results even though nothing meaningful exists, and peer review may not always catch these instances of false-positive findings. Readers can look for whether authors made analytic choices based on previous literature or a priori hypotheses, rather than seeming to test many possibilities until one reached significance. As more authors pre-register their analytic plans in advance of data collection and journals encourage this rigorous practice, it may become easier to know which findings were truly robust.

Localization.

Perhaps the most notable limitation of ERPs is their poor spatial resolution; because of the way electricity moves and spreads through the brain, skull, and scalp, voltage at a given electrode does not simply reflect activity that is generated directly below its location. Quantifying the precise anatomical region in the brain that produces an ERP component, also known as localization, is often impossible without making several assumptions about the locations and numbers of generators of that process in the brain that are often unknown. Making claims about a certain ERP’s precise location in the brain, or differences by conditions or groups, requires an extremely high burden of proof. Even with analysis technology such as LORETA (Pascual-Marqui et al. 2002), the use of high-density electrode arrays, and using individual’s MRI scans to assist in analysis, poor localization remains a fundamental limitation of EEG and ERPs. The sources of ERP signals can be estimated these processes, but we suggest proceeding with caution regarding claims about truly “localizing” an ERP.

Lateralization.

Because localization of ERPs is so limited, researchers interested in spatial questions will sometimes test whether an effect is stronger on one side of the brain than the other, known as lateralization. Lateralization is relevant to questions related to language because language functions are typically left-hemisphere lateralized; thus, one might expect to see larger amplitudes in left vs. right hemisphere electrodes. Many studies have reported lateralization being weaker or absent in language-related disorders. Lateralization can be calculated in numerous ways such as using centroids (spatial means) of scalp maps (e.g., Maurer et al. 2009) or comparing single or groups of electrodes from left and right hemispheres (e.g., Monjauze et al. 2011). Although there is no consensus on the best way to study lateralization, it is nonetheless a valid practice in ERP research.

Considerations for Experimental Design and Interpretation

Number of trials.

Researchers face many tradeoffs in designing experiments. Ideally, many trials (often, hundreds per condition) are collected so that the cleanest possible average ERP can be achieved. However, this is not always practical for many reasons. The length of an experiment paradigm is an important consideration because children (and even sleep-deprived undergraduates) may have a hard time sitting still and maintaining attention to the experiment over long periods. On one hand, increasing the number of trials can help increase statistical power for between- and within-group analyses by improving reliability (Boudewyn et al. 2018). On the other hand, limiting the length of the paradigm may allow more participants to tolerate the duration of the paradigm. Furthermore, some ERPs may habituate or disappear entirely over time (McGee et al., 2001). There is not a clear rule or guideline of how many trials are “enough” to include as this depends on how clean the data are and how large the component is. Efforts should thus be made to determine the reliability of a given ERP, or refer to published research that has determined an appropriate number of trials (e.g., Foti et al. 2013). This is challenging, as there seems to be no clear answer for how many trials are needed to reliably measure any given ERP component (e.g., anywhere from 10 to several hundred) across all EEG systems or stimuli parameters, but the number of trials presented and used/accepted that are compared per participant group and/or condition should be reported (Boudewyn et al. 2018).

Difference across conditions and motor movements.

Motor responses are another important consideration in the design of ERP experiments. For example, if the goal of an experiment is to have participants actively categorize a rare stimulus and researchers want to be sure that this is being done accurately, they may ask participants to press a button to each rare stimulus. Importantly, the ERP response to the rare stimulus will also contain the motor activity associated with the button press. This underscores an important feature of design that readers can again watch for; everything possible should be the same between two types of trials being compared. If trials in one condition have a button press (or other response), the comparison trials should as well, or the source of any differences between trials cannot be determined. Some potential solutions would be to have a button press on every trial (one button for rare and one for frequent, pressed using the same hand or finger), to add a third type of trial that is excluded from analysis in which participants press a button to indicate their attention, or to have no button press at all. Closely matched conditions thus are necessary to isolate the activity of interest with no confounding factors.

Interpreting peaks and components.

Even though components follow the general pattern of timing of different processes in the brain, they must be interpreted carefully. A common misconception is that the peak (highest or lowest point between onset and offset) indicates something meaningful about the component or underlying brain processes (Luck, 2014). The same observed waveform (that is, the plot of voltage change over time) could be made up of multiple sets of components that overlap, so the final waveform peak’s amplitude (voltage at the peak) and latency (timing at the peak) may not necessarily be important (figure 2). This is important to keep in mind; for example, it is incorrect to assume that because you see the peak of an ERP in Condition A earlier than the peak in Condition B, that this is because something about Condition A causes the component to happen earlier. There could be other components’ onsets or offsets occurring simultaneously that would shift the peaks despite the component of interest remaining stable.

Developmental changes.

A particular challenge for pediatric and longitudinal ERP research is that components change over the course of development. A pattern of auditory responses, the P1-N1-P2 complex, is very well established and studied in adults, is not found even by age 9; in children, P1 is the most prominent of these components, whereas N1 is most prominent in adults (Čeponiené et al. 2002). Indeed, there is considerable change in the morphology (i.e., shape) and timing of these and other components from infancy to childhood (Čeponiené et al. 2002, Wunderlich et al. 2006). The MMN can actually be a positive rather than a negative-going wave in infants, and is thus sometimes called the mismatch response (MMR) (van Leeuwen et al. 2006). This can make comparisons between different age groups very difficult when discussing each component on its own, especially considering the problem presented in figure 2 of overlapping components that affect each other’s shapes. It is also typical for the overall amplitude of ERPs to decrease with age. Readers should consider these factors when studies compare across ages.

Why use ERPs?

Given the complexities and challenges of ERP research, what are the advantages of this technique? Perhaps the clearest reason to use ERPs is that ERPs and behavior do not have a one-to-one relationship; if they did, there would be no reason to study ERPs when behavior could be used. For example, the MMN may be a more sensitive indicator of auditory discrimination than behavior (Stoodley et al. 2006). This means that we are measuring something unique with ERPs, that can add information to our understanding of typical and atypical behaviors. Here, we discuss the primary reasons researchers use ERPs.

When Behavior is Hard to Obtain

One of the best reasons to use ERPs is when behavioral paradigms would be ill-suited or difficult to obtain from the population of interest. A clear example of this is with auditory discrimination among infants. The mismatch response (MMR) ERP, an infant corollary of the MMN, is elicited by an oddball paradigm and does not require the attention of the infant (van Leeuwen et al. 2006). Like the MMN, the component is believed to represent an objective measure of auditory discrimination, which roughly corresponds to behavioral discrimination thresholds in adults (Näätänen and Kreegipuu 2011). As behavioral tests cannot otherwise assess this skill in very young infants, ERPs offer a window into how this process develops over time. Furthermore, as behavioral auditory discrimination is impaired in some (but not all) individuals with pediatric language delays and disorders (Benasich and Tallal 2002), one might expect this difference to be present in the brain from birth.

As A Biomarker When Behavior Can’t Tell Us Enough

Another major interest of many researchers is whether ERPs could practically aid in clinical screening, diagnosis, or progress monitoring for individuals with a disorder. An ERP used in this way would be considered a “biomarker”; a biomarker is a biological substance or process that is objectively quantifiable, reproducible, and related to underlying biology implicated in a disorder. Because so many disorders in the field of pediatric communication disorders, such as autism, language disorder, and dyslexia cannot be diagnosed until children fail to achieve typical milestones, early and objective neural markers of these disorders have been a focus of much research. Using ERPs in this way has the potential to be an extremely powerful tool, but it requires highly reliable measures. Many studies have found ERP components that relate to specific language impairment/developmental language disorder (e.g., Friedrich et al. 2004) or dyslexia (Bruder et al. 2011, Halliday et al. 2014, Maurer et al. 2009). Researchers have also measured the MMR at birth in at-risk individuals and followed them through to school age to see how well ERPs predicted who would develop dyslexia (Jyväskylä Longitudinal Study of Dyslexia; e.g., Guttorm et al. 2010). Associations with later language outcomes from these early measures show some promise, but much more research is needed to determine the generalizability of these results (see Norton et al. 2019).

To date, these biomarker studies have yet to be taken beyond the initial stages of discovery to more clinically meaningful statistical validation and implementation. Once preliminary associations are found, candidate ERP biomarkers must be subjected to cross-validation to test their generalizability and controlled replication studies (for a model of bringing brain measures to clinical utility, see Gabrieli et al. 2015). Despite these challenges, ERPs remain an enticing biomarker because they are relatively fast, easy, and inexpensive to obtain (compared to say, a detailed cognitive and language assessment or an MRI), and can be obtained early in life and thus lead to earlier identification and intervention.

Etiology of Disorders

Most language disorders are neurologically based, but their causes are not fully understood; there is hope that ERPs may be able to help elucidate causes of disorders, rather than just understand their behavioral symptoms. Even if we cannot tell much about an individual due to low reliability of measures, we may still be able to understand the nature of neural processing in groups of people with language differences or disorders with large samples and rigorous practices.

For example, one hypothesis about the cause of dyslexia is that there is a neural impairment in the way that sounds are processed and discriminated. Comparing the MMN ERPs from two groups that are carefully matched (e.g., in age, number of included trials, etc.) can tell us whether this automatic auditory process is altered in language disorders. All studies, regardless of the hypothesis and ERP components used, should discuss how their results tested their hypothesis as explicitly as possible.

Furthering Basic Understanding of Neural Processing

ERPs have also been applied at a more basic level to understand how humans process different types of information. The semantic incongruity paradigm has been used extensively to describe how the brain processes linguistic information. Other examples of basic applications of ERPs are how the brain form auditory objects (e.g., Sussman and Steinschneider 2009), how it integrates information from multiple senses (Brett-Green et al. 2008), and how it processes emotion in faces (Blau et al. 2007). Because these uses aren’t particularly applicable clinically, we won’t discuss them further, but many of the applied uses of ERPs were born out of these basic research applications.

When and Why Not to Use ERPs

Despite the advantages of ERPs in many situations, there are also several situations where it is most appropriate to not use this technique. When a behavioral experiment can answer the same question, there is often no reason to “tack on” ERPs. That is, just because a process can be measured using an ERP doesn’t mean it should; a study should use ERPs when they are the right tool to answer a given question. If the research questions focus on where a process occurs in the brain, ERPs may not be a good technique relative to other methods such as fMRI or MEG. ERP studies are less well-suited to certain populations, such as those using cochlear implants, as the electrical activity and placement of the implant receiver may affect the signal.

Clinical Implications and Conclusions

In our view, research using ERPs is promising, but has not yet made a direct impact on how clinical diagnosis or treatment of speech or language disorders proceeds. Although ERPs have been used in basic research to understand cognitive processes for decades, their applications to clinical questions has been growing as technology becomes more affordable and accessible. Using ERPs to understand clinical disorders is still somewhat in its infancy, and it is important to recognize that ERP research overall has evolved and been refined over decades. It will likely take decades more of rigorous research to bring ERPs to clinical application, as we learn more every year about the most valid and reliable ways to measure ERPs and technology makes them more feasible to collect (including that they are more participant-friendly, quick to set up, and affordable). At the same time, we are also learning more about the heterogeneity of speech and language disorders, and how we can best address this heterogeneity both clinically and scientifically. Nonetheless, we believe it is important for clinicians to interpret ERP results (i) because informed readers will push researchers to improve the state of science, (ii) because understanding the shortcomings in existing research can provide a road map of what to look for in the years to come in ERP experiments, and (iii) lastly, if ERPs move into clinical practice in the future, clinicians can be ready to apply their knowledge of the data to their patients as if it were standardized behavioral assessment data. Although ERPs are not used in the clinic to assess speech and language disorders at this time, we believe that scientific literacy on the topic is key to facilitating research that would help achieve this goal.

What this paper adds:

Event-related potentials (ERPs) have become more prevalent in research relevant to communication sciences and disorders. In order for clinicians to review and evaluate this research, an understanding of ERPs is needed. This review adds to the field by providing an accessible description of what ERPs are, description of what ERP components are and the most relevant commonly-used components, as well as how ERP data are recorded and processed. With this foundational understanding of how ERPs work, guidelines for interpretation of ERP data are given. Though few ERP studies right now have direct implications for clinical practice, we discuss several ways through which ERPs can impact clinical practice in the future, through providing information that cannot be obtained by behavior alone about the etiology of disorders, and as potential biomarkers of disorder or treatment response.

Acknowledgments

We acknowledge Steve Luck, whose writing and teaching has guided many of the topics covered in this tutorial. We thank Katherine Wolfert for comments on EEG net setup and Brittany Manning CCC-SLP for discussion on clinical considerations. E.S.N. was supported by the National Institute on Deafness and Other Communication Disorders (NIDCD) under award number R01DC016273. The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health.

References

  1. Allison T, Matsumiya Y, Goff GD, and Goff WR, 1977, The scalp topography of human visual evoked potentials. Electroencephalography and Clinical Neurophysiology, 42, 185–197. 10.1016/0013-4694(77)90025-6. [DOI] [PubMed] [Google Scholar]
  2. Bell MA and Cuevas K, 2012, Using EEG to study cognitive development: Issues and practices. Journal of Cognition and Development, 13, 281–294. 10.1080/15248372.2012.691143. [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Benasich AA and Tallal P, 2002, Infant discrimination of rapid auditory cues predicts later language impairment. Behavioural Brain Research, 136, 31–49. 10.1016/S0166-4328(02)00098-0. [DOI] [PubMed] [Google Scholar]
  4. Bishop DVM, 2007, Using mismatch negativity to study central auditory processing in developmental language and literacy impairments: Where are we, and where should we be going? Psychological Bulletin, 133, 651–672. 10.1037/0033-2909.133.4.651. [DOI] [PubMed] [Google Scholar]
  5. Blau VC, Maurer U, Tottenham N, and McCandliss BD, 2007, The face-specific N170 component is modulated by emotional facial expression. Behavioral and Brain Functions, 3, 7 10.1186/1744-9081-3-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Boudewyn MA, Luck SJ, Farrens JL, and Kappenman ES, 2018, How many trials does it take to get a significant ERP effect? It depends. Psychophysiology, 55, e13049 10.1111/psyp.13049. [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Brett-Green BA, Miller LJ, Gavin WJ, and Davies PL, 2008, Multisensory integration in children: A preliminary ERP study. Brain Research, 1242, 283–290. 10.1016/j.brainres.2008.03.090. [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Brouwer A-M and Van Erp JBF, 2010, A tactile P300 brain-computer interface. Frontiers in Neuroscience, 4:19 10.3389/fnins.2010.00019. [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Bruder J, Leppänen PHT, Bartling J, Csépe V, Démonet J-F, and Schulte‐Körne G, 2011, Children with dyslexia reveal abnormal native language representations: Evidence from a study of mismatch negativity. Psychophysiology, 48, 1107–1118. 10.1111/j.1469-8986.2011.01179.x. [DOI] [PubMed] [Google Scholar]
  10. Buzsáki G, Anastassiou CA, and Koch C, 2012, The origin of extracellular fields and currents — EEG, ECoG, LFP and spikes. Nature reviews neuroscience, 13, 407–420. 10.1038/nrn3241. [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Cantiani C, Choudhury NA, Yu YH, Shafer VL, Schwartz RG, and Benasich AA, 2016, From Sensory Perception to Lexical-Semantic Processing: An ERP Study in Non-Verbal Children with Autism. PLOS ONE, 11, e0161637 10.1371/journal.pone.0161637. [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Čeponiené R, Rinne T, and Näätänen R, 2002, Maturation of cortical sound processing as indexed by event-related potentials. Clinical Neurophysiology, 113, 870–882. 10.1016/S1388-2457(02)00078-0. [DOI] [PubMed] [Google Scholar]
  13. Deveney CM, Briggs-Gowan MJ, Pagliaccio D, Estabrook CR, Zobel E, Burns JL, Norton ES, Pine DS, Brotman MA, Leibenluft E, and Wakschlag LS, 2019, Temporally sensitive neural measures of inhibition in preschool children across a spectrum of irritability. Developmental Psychobiology, 61, 216–227. 10.1002/dev.21792. [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Foti D, Kotov R, and Hajcak G, 2013, Psychometric considerations in using error-related brain activity as a biomarker in psychotic disorders. Journal of Abnormal Psychology, 122, 520–531. 10.1037/a0032618. [DOI] [PubMed] [Google Scholar]
  15. Friedrich M, Weber C, and Friederici AD, 2004, Electrophysiological evidence for delayed mismatch response in infants at-risk for specific language impairment. Psychophysiology, 41, 772–782. 10.1111/j.1469-8986.2004.00202.x. [DOI] [PubMed] [Google Scholar]
  16. Gabrieli JDE, Ghosh SS, and Whitfield-Gabrieli S, 2015, Prediction as a humanitarian and pragmatic contribution from human cognitive neuroscience. Neuron, 85, 11–26. 10.1016/j.neuron.2014.10.047. [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Guttorm TK, Leppänen PHT, Hämäläinen JA, Eklund KM, and Lyytinen HJ, 2010, Newborn event-related potentials predict poorer pre-reading skills in children at risk for dyslexia. Journal of Learning Disabilities, 43, 391–401. 10.1177/0022219409345005. [DOI] [PubMed] [Google Scholar]
  18. Halliday LF, Barry JG, Hardiman MJ, and Bishop DV, 2014, Late, not early mismatch responses to changes in frequency are reduced or deviant in children with dyslexia: an event-related potential study. Journal of Neurodevelopmental Disorders, 6, 21 10.1186/1866-1955-6-21. [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. He JB, 2015, Electrocorticogram (ECoG) In Jaeger D and Jung R (eds), Encyclopedia of Computational Neuroscience ((New York, NY: Springer Publishing Company; )), pp. 1070–1074. [Google Scholar]
  20. Henderson LM, Baseler HA, Clarke PJ, Watson S, and Snowling MJ, 2011, The N400 effect in children: Relationships with comprehension, vocabulary and decoding. Brain and Language, 117, 88–99. 10.1016/j.bandl.2010.12.003. [DOI] [PubMed] [Google Scholar]
  21. Homan RW, Herman J, and Purdy P, 1987, Cerebral location of international 10–20 system electrode placement. Electroencephalography and Clinical Neurophysiology, 66, 376–382. 10.1016/0013-4694(87)90206-9. [DOI] [PubMed] [Google Scholar]
  22. Keil A, Debener S, Gratton G, Junghöfer M, Kappenman ES, Luck SJ, Luu P, Miller GA, and Yee CM, 2014, Committee report: Publication guidelines and recommendations for studies using electroencephalography and magnetoencephalography. Psychophysiology, 51, 1–21. 10.1111/psyp.12147. [DOI] [PubMed] [Google Scholar]
  23. Kiesel A, Miller J, Jolicœur P, and Brisson B, 2008, Measurement of ERP latency differences: A comparison of single-participant and jackknife-based scoring methods. Psychophysiology, 45, 250–274. 10.1111/j.1469-8986.2007.00618.x. [DOI] [PubMed] [Google Scholar]
  24. Kimura M, Katayama J, Ohira H, and Schröger E, 2009, Visual mismatch negativity: New evidence from the equiprobable paradigm. Psychophysiology, 46, 402–409. 10.1111/j.1469-8986.2008.00767.x. [DOI] [PubMed] [Google Scholar]
  25. Kutas M and Hillyard SA, 1980, Reading senseless sentences: brain potentials reflect semantic incongruity. Science, 207, 203–205. 10.1126/science.7350657. [DOI] [PubMed] [Google Scholar]
  26. Kutas M and Hillyard SA, 1984, Brain potentials during reading reflect word expectancy and semantic association. Nature, 307, 161–163. 10.1038/307161a0. [DOI] [PubMed] [Google Scholar]
  27. van Leeuwen T, Been P, Kuijpers C, Zwarts F, Maassen B, and van der Leij A, 2006, Mismatch response is absent in 2-month-old infants at risk for dyslexia. NeuroReport, 17, 351 10.1097/01.wnr.0000203624.02082.2d. [DOI] [PubMed] [Google Scholar]
  28. Loken E and Gelman A, 2017, Measurement error and the replication crisis. Science, 355, 584–585. 10.1126/science.aal3618. [DOI] [PubMed] [Google Scholar]
  29. Luck SJ, 2005, Ten simple rules for designing ERP experiments In Handy TC (ed), Event-related potentials: a methods handbook ((Cambridge, MA: MIT Press; )), pp. 17–32. [Google Scholar]
  30. Luck SJ, 2014, An introduction to the event-related potential technique. 2nd ed. (Cambridge, MA: MIT Press; ). [Google Scholar]
  31. Luck SJ and Gaspelin N, 2017, How to get statistically significant effects in any ERP experiment (and why you shouldn’t). Psychophysiology, 54, 146–157. 10.1111/psyp.12639. [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Luck SJ and Kappenman ES, 2011, The Oxford handbook of event-related cotential components (Oxford, England: Oxford University Press; ). [Google Scholar]
  33. Luck SJ, Vogel EK, and Shapiro KL, 1996, Word meanings can be accessed but not reported during the attentional blink. Nature, 383, 616–618. 10.1038/383616a0. [DOI] [PubMed] [Google Scholar]
  34. Maurer U, Bucher K, Brem S, Benz R, Kranz F, Schulz E, van der Mark S, Steinhausen H-C, and Brandeis D, 2009, Neurophysiology in preschool improves behavioral prediction of reading ability throughout primary school. Biological Psychiatry, 66, 341–348. 10.1016/j.biopsych.2009.02.031. [DOI] [PubMed] [Google Scholar]
  35. McCallum WC, Farmer SF, and Pocock PV, 1984, The effects of physical and semantic incongruities on auditory event-related potentials. Electroencephalography and Clinical Neurophysiology/Evoked Potentials Section, 59, 477–488. 10.1016/0168-5597(84)90006-6. [DOI] [PubMed] [Google Scholar]
  36. McGee TJ, King C, Tremblay K, Nicol TG, Cunningham J, and Kraus N, 2001, Long-term habituation of the speech-elicited mismatch negativity. Psychophysiology, 38, 653–658. [PubMed] [Google Scholar]
  37. Monjauze C, Broadbent H, Boyd SG, Neville BGR, and Baldeweg T, 2011, Language deficits and altered hemispheric lateralization in young people in remission from BECTS. Epilepsia, 52, e79–e83. 10.1111/j.1528-1167.2011.03105.x. [DOI] [PubMed] [Google Scholar]
  38. Näätänen R and Kreegipuu K, 2011, The Mismatch Negativity (MMN) In The Oxford handbook of event-related potential components ((Oxford University Press; )). [Google Scholar]
  39. National Center for Hearing Assessment and Management, 2011, State early hearing detection and intervention/universal newborn hearing screening mandates (Utah State University; ). [Google Scholar]
  40. Norton ES, Gabrieli JDE, and Gaab N, 2019, Neural predictors of dyslexia In Verhoeven L, Perfetti C, and Pugh K (eds), Developmental Dyslexia across Languages and Writing Systems: A Handbook ((Cambridge University Press; )), pp. 253–276. [Google Scholar]
  41. Pascual-Marqui RD, Esslen M, Kochi K, and Lehmann D, 2002, Functional imaging with low resolution brain electromagnetic tomography (LORETA): a review. Clinical Pharmacology, 12. [PubMed] [Google Scholar]
  42. Plakas A, van Zuijen T, van Leeuwen T, Thomson JM, and van der Leij A, 2013, Impaired non-speech auditory processing at a pre-reading age is a risk-factor for dyslexia but not a predictor: An ERP study. Cortex, 49, 1034–1045. 10.1016/j.cortex.2012.02.013. [DOI] [PubMed] [Google Scholar]
  43. Polich J, 2007, Updating P300: An integrative theory of P3a and P3b. Clinical Neurophysiology, 118, 2128–2148. 10.1016/j.clinph.2007.04.019. [DOI] [PMC free article] [PubMed] [Google Scholar]
  44. Polich J, 2011, Neuropsychology of P300 In Luck SJ and Kappenman ES (eds), The Oxford handbook of event-related potential components ((Oxford, England: Oxford University Press; )), pp. 159–188. [Google Scholar]
  45. Pratt H, 2011, Sensory ERP components In Luck SJ and Kappenman ES (eds), The Oxford handbook of event-related potential components ((Oxford University Press; )), pp. 89–114. [Google Scholar]
  46. Puce A and Hämäläinen MS, 2017, A review of issues related to data acquisition and analysis in EEG/MEG studies. Brain Sciences, 7, 58 10.3390/brainsci7060058. [DOI] [PMC free article] [PubMed] [Google Scholar]
  47. Sanders LD, Newport EL, and Neville HJ, 2002, Segmenting nonsense: an event-related potential index of perceived onsets in continuous speech. Nature Neuroscience, 5, 700–703. 10.1038/nn873. [DOI] [PMC free article] [PubMed] [Google Scholar]
  48. Skoe E and Kraus N, 2010, Auditory brainstem response to complex sounds: a tutorial. Ear and hearing, 31, 302–324. 10.1097/AUD.0b013e3181cdb272. [DOI] [PMC free article] [PubMed] [Google Scholar]
  49. Stoodley CJ, Hill PR, Stein JF, and Bishop DVM, 2006, Auditory event-related potentials differ in dyslexics even when auditory psychophysical performance is normal. Brain Research, 1121, 190–199. 10.1016/j.brainres.2006.08.095. [DOI] [PubMed] [Google Scholar]
  50. Sussman E and Steinschneider M, 2009, Attention effects on auditory scene analysis in children. Neuropsychologia, 47, 771–785. 10.1016/j.neuropsychologia.2008.12.007. [DOI] [PMC free article] [PubMed] [Google Scholar]
  51. Tanner D, Morgan-Short K, and Luck SJ, 2015, How inappropriate high-pass filters can produce artifactual effects and incorrect conclusions in ERP studies of language and cognition. Psychophysiology, 52, 997–1009. 10.1111/psyp.12437. [DOI] [PMC free article] [PubMed] [Google Scholar]
  52. Weiss S and Mueller HM, 2003, The contribution of EEG coherence to the investigation of language. Brain and Language, 85, 325–343. 10.1016/S0093-934X(03)00067-1. [DOI] [PubMed] [Google Scholar]
  53. Wunderlich JL, Cone-Wesson BK, and Shepherd R, 2006, Maturation of the cortical auditory evoked potential in infants and young children. Hearing Research, 212, 185–202. 10.1016/j.heares.2005.11.010. [DOI] [PubMed] [Google Scholar]

RESOURCES