Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2018 May 1.
Published in final edited form as: Psychophysiology. 2017 Feb 13;54(5):780–791. doi: 10.1111/psyp.12827

Evaluating the efficacy of fully automated approaches for the selection of eye blink ICA components

Matthew B Pontifex 1, Vladimir Miskovic 2, Sarah Laszlo 2
PMCID: PMC5397386  NIHMSID: NIHMS840363  PMID: 28191627

Abstract

Independent component analysis (ICA) offers a powerful approach for the isolation and removal of eye blink artifacts from EEG signals. Manual identification of the eye blink ICA component by inspection of scalp map projections, however, is prone to error, particularly when non-artifactual components exhibit topographic distributions similar to the blink. The aim of the present investigation was to determine the extent to which automated approaches for selecting eye blink related ICA components could be utilized to replace manual selection. We evaluated popular blink selection methods relying on spatial features [EyeCatch()], combined stereotypical spatial and temporal features [ADJUST()], and a novel method relying on time-series features alone [icablinkmetrics()] using both simulated and real EEG data. The results of this investigation suggest that all three methods of automatic component selection are able to accurately identify eye blink related ICA components at or above the level of trained human observers. However, icablinkmetrics(), in particular, appears to provide an effective means of automating ICA artifact rejection while at the same time eliminating human errors inevitable during manual component selection and false positive component identifications common in other automated approaches. Based upon these findings, best practices for 1) identifying artifactual components via automated means and 2) reducing the accidental removal of signal-related ICA components are discussed.

Keywords: Independent Component Analysis, EEG Artifact, EEGLAB


Over the past decade an increasing number of laboratories have begun to utilize temporal independent component analysis (ICA) to isolate and remove eye blink artifacts from EEG signals. Indeed, examination of the journal Psychophysiology over the past two years reveals that nearly the same proportion of EEG investigations utilize temporal ICA approaches for eye blink artifact reduction/correction as those investigations that utilize regression-based approaches. This is undoubtedly related in some part to the growing adoption of EEGLAB (Delorme & Makeig, 2004), a MATLAB / Octave based graphical toolbox for data processing, which implements temporal ICA in its standard workflow. Of practical importance, this temporal ICA approach to artifact correction may not be globally appropriate for all artifacts — such as saccadic eye movements and non-stationary artifacts, (see Hoffmann & Falkenstein, 2008 for further discussion). However, for eye blink artifact correction, the ICA approach has been found to exhibit superior performance relative to regression-based approaches (Jung, et al., 2000). An important distinction to make, however, is that unlike regression-based approaches, ICA does not inherently perform any form of artifact correction. That is, these approaches are simply blind-source signal separation techniques that attempt to dissociate temporally independent yet spatially fixed components (Bell & Sejnowski, 1995). Thus, a critical limitation of temporal ICA-based approaches to artifact correction is the reliance on subjective human judgments to determine what components are associated with noise rather than signal, so that the data can be back-projected to reconstruct EEG signals in the absence of artifactual activity. Although automated approaches exist, we have little understanding of the extent to which these automated ICA component selection approaches are robust to variation in signal-to noise-ratio or across varying electrode densities. Thus, the aim of the present investigation was to determine if fully automated approaches for selecting eye blink related ICA components can and should be utilized to replace manual selection of eye blink artifact components by human users.

In a common EEGLAB workflow, following separation of the signals using standard ICA algorithms, a human observer must visually sift through the full set of temporal ICA components in order to manually select one or more components for removal. Such an approach is not only labor intensive, but it is also user-dependent, making it more prone to errors, or, potentially, to bias (e.g., quality control is dependent on the user’s expertise level). Human fallibility is especially relevant in the case of experiments concerned with frontal ERP / EEG activity, where it is particularly difficult to differentiate the scalp projection of the temporal ICA component(s) associated with eye blink activity from the scalp projection of the temporal ICA component(s) associated with genuine, frontally maximal cortical activity, such as the LAN (Castellanos & Makarov, 2006). Although it is possible to inform these decisions by inspecting temporal ICA activations (see Figure 1), such approaches are not explicitly detailed within the EEGLAB documentation, rendering knowledge and implementation of such methodologies potentially variable across research laboratories.

Figure 1.

Figure 1

Illustration of eye blinks recorded from the VEOG electrode, how they manifest across three midline electrode sites, and how the eye blink related signal is separated from other signals using ICA decomposition. Note the high degree of similarity between the VEOG electrode and ICA component 1.

To address the potential limitations of human-selected ICA artifact rejection, a number of methods for automatic identification of artifact related temporal ICA components have been developed for EEGLAB. These include ADJUST (Mognon, Jovicich, Bruzzone, & Buiatti, 2011), CORRMAP (Viola, Thorne, Edmonds, Schneider, & Eichele, 2009), and EyeCatch (Bigdely-Shamlo, Kreutz-Delgado, Kothe, & Makeig, 2013). It should be pointed out that such approaches are simply added on following the application of ICA to the data in order to automate the selection of artifact related components, and are not used as a replacement for the ICA application. The most widely downloaded software plugin is the ADJUST() method, which attempts to identify a wide array of potential sources of artifact such as eye blinks, eye movements, cardiac induced artifacts, and other stereotypical movements. To this end, temporal independent components are characterized by combining both spatial and temporal information, with the identification of artifactual components based on stereotyped spatio-temporal features such as temporal kurtosis and the spatial average difference. Alternatively, the EyeCatch() plugin attempts to distinguish temporal ICA components based on the correlation between the scalp map projection for each ICA component and a database of 3,452 (as of the writing of this manuscript) exemplar eye-activity related template scalp maps. This approach has been found to exhibit overall performance similar to the CORRMAP() function (Viola, Thorne, Edmonds, Schneider, & Eichele, 2009) but has an advantage over CORRMAP() in that it is fully automated.

A limitation of the approaches mentioned above is that they largely rely on ancillary indices of ICA components, such as scalp map projections of temporal ICA weights, to differentiate eye-blink related components from non-blink related components, rendering them prone to the same potential sources of error as the human visual inspection approach — particularly when non-artifactual frontally maximal ICA components occur. Since these approaches rely on the scalp topography of the temporal ICA components, the eye-blink component can easily be confused with non-artifactual frontally distributed components. This is a fundamental weakness of any topography-based approach to selecting the eye blink related temporal ICA component. In contrast, a time-domain approach should not be vulnerable to potential confusion between frontally distributed non-artifact components and eye blink related components. Accordingly, we were interested in comparing a time-domain approach to the existing spatial approaches. Consequently, we developed a time-domain approach, icablinkmetrics(), predicated on two basic premises: 1) that the temporal ICA component(s) associated with eye blinks should be related to the eye blink activity present within the EEG more so than any other temporal ICA component (e.g., via correlation and convolution), and 2) that removal of the temporal ICA component associated with the eye blinks should reduce the eye blink artifact present within the EEG data more so than the removal of any other temporal ICA component following back projection (when the data is reconstructed without the artifactual component). Again this approach is simply added on following the application of ICA to the data in order to automate the selection of artifact related components, and is not used as a replacement to the ICA application. For a more detailed description of the background and theory underlying the premises which guided the development of the icablinkmetrics() time-domain approach see the Supplementary Appendix. In the interest of transparency, we note that the icablinkmetrics() approach was created by M. P. with input from S. L. The icablinkmetrics() EEGLAB plugin—which can be run from either the command line or the Tools menu of EEGLAB — is available through the EEGLAB Extension Manager or by downloading from http://sccn.ucsd.edu/wiki/EEGLAB_Extensions.

The aim of the present investigation was to assess the efficacy of these automated approaches for the selection of eye blink related artifact components. To this end, we evaluated the relative merits of automatic approaches to eye-blink component selection methods relying on time-series data [icablinkmetrics()] as compared to those relying on combined stereotypical spatial and temporal features [ADJUST(), (Mognon, Jovicich, Bruzzone, & Buiatti, 2011)], or spatial features alone [EyeCatch(), (Bigdely-Shamlo, Kreutz-Delgado, Kothe, & Makeig, 2013)]. An intrinsic weakness of a temporal approach to eye blink related component selection is that with increasing noise in the time series, procedures for identifying when eye blink related activity occurs are more prone to failure. To examine this issue, we utilized simulated EEG data to investigate the extent to which each of these automated approaches would be sensitive to variation in the magnitude of the eye blink artifact amid increasing levels of noise in the signal. Next, we assessed the generalizability of these automated approaches across real EEG data collected with varying electrode densities and in response to different tasks. Finally, for comparison, we assessed the accuracy of the current common method of trained observers visually selecting temporal ICA components. Collectively, these analyses serve to address the critical question of whether fully automated approaches for selecting eye blink related ICA components can and should be utilized to replace manual selection of eye artifact components by human users, and, if so, what their potential vulnerabilities are.

Simulated EEG Varying in the Magnitude of the Artifact and Level of Noise

Method

A total of 3,072 simulated EEG datasets were created matched to three (real) exemplar EEG datasets (1,024 simulations per exemplar dataset). For each exemplar dataset, simulated data were created representing a wide range of possible eye blink artifact magnitude and noise conditions. In this context, the aim was not to simulate the computational processes by which the EEG signal is actually created in the brain (e.g., Laszlo & Armstrong, 2014; Laszlo & Plaut, 2012). Rather, our goal was to ensure that the artificial data exhibited the same frequency domain properties and signal-to-noise ratio (prior to the injection of more noise per the experimental manipulations) in the same amplitude range as true EEG data. Such an approach enabled the creation of EEG datasets that had similar properties to real EEG, while allowing for the ability to modulate the level of noise present within the signal as a function of the variability found within real EEG datasets. Simulated EEG datasets were created by 1) Fourier decomposing each exemplar dataset at each channel and then 2) producing weighted sums of sines with random phase shifts that resulted in simulated datasets with the same frequency characteristics as the exemplar. The first and last 100 points of the simulated time series were removed to account for edge artifacts from the finite sum of sines, and simulated time series were scaled to have the same mean and standard deviation as the exemplar datasets, per channel. Each simulated dataset contained 25,480 points for each of 28 channels, allowing 32.5 data points for each ICA weight (data points/channels2). Noise was added to the simulated datasets by randomly perturbing both the phase and amplitude at each point in the time series. Phase perturbations were distributed uniformly, amplitude perturbations were distributed normally. The noise perturbations within the simulated EEG data were scaled to create 32 levels of noise ranging from 0.4 to 10 times the standard deviation of the exemplar EEG dataset in increments of 0.31 standard deviations. Simulated data constructed in this manner do not include eye blink artifacts, and thus constitute the “ground truth” for ICA artifact correction. That is, this data can be compared with reconstructed data created by removing each ICA component. The reconstructed data that is most similar to the ground truth data must then reflect removal of the truly artifactual eye blink component (as opposed to the other, non-artifactual components).

Eye blink artifacts were then introduced into the simulated data using a Chebyshev window (250 ms in length) as the model eye blink. Twenty eye blinks were introduced into the simulated time series at a rate of roughly one blink every 1.25 seconds with the propagation of the simulated blinks across the scalp controlled by a spherical head model derived empirically from the exemplar EEG dataset. The simulated eye blinks were scaled to create 32 levels of artifact magnitude ranging from 20 to 300 µV in increments of 9 µV. This approach therefore allowed for the examination of the automated eye blink component selection algorithms across an extreme range of signal to noise ratios. Figure 2 provides exemplars of the simulated EEG across the range of possible eye blink artifact magnitude and noise conditions.

Figure 2.

Figure 2

Representative data illustrating the simulated EEG across the range of possible eye blink artifact magnitude and noise conditions for three electrode sites. For reference, the time points for the seeded eye blinks are highlighted in green.

Following each simulation, ICA decompositions were performed using the extended Infomax algorithm to extract subgaussian components using the default settings called for the binary instance of this function in EEGLAB. To identify the components related to the simulated artifact, the mean difference (as an absolute value) between the blink-free simulated data and the reconstructed simulated data was computed following back projection of the data without each ICA component, separately. As the eye blink component(s) should be rare relative to the other components, the truly artifactual components were selected by normalizing the differences, and computing the probability of the difference occurring given a normal distribution (see Figure 3). Those components with a probability less than 0.05 were identified as truly artifactual components. Across the 3,072 simulations, the truly artifactual ICA component was identified in 1,700 (55.3%) of the simulations with instances where the truly artifactual component was unable to be determined occurring when the magnitude of the noise far exceeded the magnitude of the eye blink (see Figures 2 and 4). Comparison of the automated component selection procedures was restricted to only those simulations where the truly artifactual component was able to be identified.

Figure 3.

Figure 3

Representative data illustrating how the ground truth artifact-related ICA component was identified in the simulated EEG data. Only the removal of a single component returns the simulated data to near its uncontaminated state, with the normalized difference between the uncontaminated data and the contaminated data following removal of the ICA component reflecting that component as an outlier. As most components should be unrelated to the artifact, any component identified as an outlier was considered as related to the artifact.

Figure 4.

Figure 4

Graphic illustration of the results of 3,072 simulations of EEG data (1,024 simulations per exemplar dataset) for the likelihood of identifying the artifact (sensitivity) and the likelihood of misidentifying signal as artifact (1-specificity) as a function of eye blink magnitude and noise for each automated procedure. As each exemplar dataset was used to test the full range of signal to noise, some data points may only reflect a singular simulation whereas others may reflect the result of three simulations at that eye blink magnitude and noise level. Areas where the ground truth eye blink component was unable to be determined (occurring in 1,372 of the 3,072) are uncolored.

Each of the three automated procedures (icablinkmetrics() version 3.1, ADJUST() version 1.1.1, and EyeCatch()) was then tested using their default parameters. The icablinkmetrics() function was run using the VEOG channel of the simulated dataset as the artifact comparison channel. The icablinkmetrics() function identified eye blinks within the artifact channel by cross-correlating a canonical eye blink waveform using the eyeblinklatencies() function, only accepting seeded eye blinks which exhibited correlations of 0.96 or higher. Quantification of the efficacy of the automated component selection approaches for reducing the simulated artifact was performed by computing the percent reduction in the difference between the blink-free simulated data and the reconstructed simulated data ([absolute value([difference between data with simulated eye blink and blink-free data] − [difference between reconstructed data following artifact removal and blink-free data]) / (difference between data with simulated eye blink and blink-free data)]; see Table 1). Perfect reconstruction of the simulated data to its blink-free state would thus be reflected by 100% reduction in the difference between the blink-free simulated data and the reconstructed simulated data following artifact removal. All data processing was conducted using an Apple iMac with a 3.5 GHz Intel Core i7 processor and 32 GB of 1600 MHz DDR3 SDRAM.

Table 1.

ICA component classifications

True Positive True Negative False Positive False Negative Sensitivity Specificity Reduction of
Artifact
Eye blink correctly
classified
Non blink correctly
classified
Said it was eye
blink but it was not
Said it was not an
eye blink but it was
TP /
(TP+FN)
TN /
(TN+FP)
Based on
components
selected
(Rejected Artifact) (Retained Signal) (Rejected Signal) (Retained Artifact) (Identify Artifact) (Identify Signal)
Simulated Data
  icablinkmetrics() 1234 45900 0 466 72.6% 100% 89.5%
  ADJUST() 1662 45436 464 38 97.8% 99.0% 82.8%
  EyeCatch() 1560 45428 472 140 91.8% 99.0% 83.9%
Real Data
  icablinkmetrics() 92 4936 0 0 100% 100% 88.0%
    32 Channel Array 40 998 0 0 100% 100% 93.4%
    64 Channel Array 38 2233 0 0 100% 100% 89.3%
    128 Channel Array 14 1705 0 0 100% 100% 69.2%
  ADJUST() 89 4738 198 3 96.7% 96.0% 86.6%
    32 Channel Array 39 959 39 1 97.5% 96.1% 91.3%
    64 Channel Array 38 2110 123 0 100% 94.5% 88.7%
    128 Channel Array 12 1669 36 2 85.7% 97.9% 67.2%
  EyeCatch() 92 4847 89 0 100% 98.2% 87.2%
    32 Channel Array 40 950 48 0 100% 95.2% 93.3%
    64 Channel Array 38 2212 21 0 100% 99.1% 89.2%
    128 Channel Array 14 1685 20 0 100% 98.8% 64.1%
  Expert Observer 89 4930 6 3 96.7% 99.9% 85.4%
    32 Channel Array 38 994 4 2 95.0% 99.6% 88.9%
    64 Channel Array 38 2233 0 0 100% 100% 89.3%
    128 Channel Array 13 1703 2 1 92.9% 99.9% 64.6%
  Competent Observer 81 4921 15 11 88.0% 99.7% 79.8%
    32 Channel Array 38 994 4 2 95.0% 99.6% 88.9%
    64 Channel Array 38 2232 1 0 100% 100% 89.3%
    128 Channel Array 5 1695 10 9 35.7% 99.4% 27.7%
  Novice Observer 81 4888 48 11 88.0% 99.0% 82.4%
    32 Channel Array 38 977 21 2 95.0% 97.9% 89.0%
    64 Channel Array 37 2230 3 1 97.4% 99.9% 88.9%
    128 Channel Array 6 1681 24 8 42.9% 98.6% 46.0%

Note: Values indicate the number of components. The values for Reduction of Artifact indicate the percentage of the artifact removed following removal of the ICA components identified as artifactual. For the simulated data, this value reflects the percent similarity between the simulated data prior to the introduction of eye blink artifacts and the reconstructed data following removal of the selected ICA components. For the real data, this value reflects the percent reduction of the convolution (i.e., overlap) between the mean eye blink artifact and the EEG activity across all electrode sites during this same period following removal of the selected ICA components.

Statistical Analysis

The efficacy of the automated procedures for identifying the eye blink ICA component were examined statistically by evaluating their sensitivity (the likelihood of correctly identifying the eye blink ICA component(s); i.e., hits) and specificity (the likelihood of correctly not identifying a non-blink component as an eye blink ICA component(s); i.e., correct rejections) relative to the truly artifactual component. As all simulated datasets were contaminated by eye blink artifact, failure to select an eye blink component was considered a false negative error (‘miss’), unless the truly artifactual component was unable to be determined (e.g., such as if the Infomax algorithm was unable to separate the seeded eye blink from the background noise).

Results

Component selection counts along with the sensitivity and specificity are provided in Table 1. A graphical illustration of the likelihood of identifying the artifact (sensitivity) and the likelihood of misidentifying signal as artifact (1-specificity) as a function of eye blink magnitude and noise for each automated procedure is provided in Figure 4. Results of the simulation indicate that icablinkmetrics() exhibited a lower sensitivity level (72.6%) than ADJUST() and EyeCatch(), which exhibited sensitivities above 91%. The sensitivity of icablinkmetrics() and EyeCatch() was observed to vary as a function of the magnitude of the eye blink artifact and the relative noise level, with both demonstrating perfect sensitivity when the artifact amplitude to noise ratio was high. However, as the artifact amplitude to noise ratio was reduced so too was the sensitivity (see Figure 4). In contrast, ADJUST() exhibited a less interpretable pattern of decreases in sensitivity.

Although icablinkmetrics() exhibited reduced sensitivity relative to the other methods, it also displayed perfect specificity (i.e., it never made any false alarms) regardless of the artifact amplitude or noise level of the simulated data. The specificity of ADJUST() was observed to vary as a function of the magnitude of the eye blink artifact and the relative noise level, demonstrating perfect specificity when the artifact amplitude to noise ratio was high. However, as the artifact amplitude to noise ratio was reduced so too was the specificity (see Figure 4). In contrast, EyeCatch() exhibited a less interpretable pattern of decreases in specificity, seeming to have a greater incidence of falsely identifying components as artefactual when the noise level was the lowest. Additionally, icablinkmetrics() was observed to exhibit a 0% false discovery rate with the removal of the selected components resulting in 89.5% similarity to the original blink-free simulated data, whereas ADJUST() and EyeCatch() were observed to exhibit false discovery rates of 21.8% and 23.2%, respectively, with removal of the selected components resulting in less than an 84% similarity to the original blink-free simulated data. However, when restricted to only those instances where all three automated component selection approaches were able to identify a component as artifactual — thereby ensuring equivalent comparisons free from potential bias related to the failure to identify a component; the components selected by icablinkmetrics(), ADJUST(), and EyeCatch() were all observed to return the data with approximately 91% similarity to the original blink-free simulated data.

Discussion

The aim of this section was to evaluate the extent to which automatic eye-blink ICA component selection methods would be sensitive to variation in the magnitude of the eye blink artifact amid increasing levels of noise in the signal. Utilizing simulated EEG data with an identifiable truly artifactual eye blink ICA component revealed that, sensibly, decreases in the ratio between the artifact amplitude and the noise appeared to negatively impact each of the automated selection approaches. For the time-series approach utilized by icablinkmetrics(), decreases in the ratio between the artifact amplitude and the noise resulted in a reduced ability to identify a component as related to the artifact. However, despite alterations in the amplitude of the artifact and the noise, icablinkmetrics() never falsely identified a non-artifactual component as related to the eye blink. Under fully automated implementations then, icablinkmetrics() might fail to identify ICA components associated with the eye blink with noisier datasets but would seem to be robust against falsely removing signal-related ICA components (i.e., it errs on the side of caution), as reflected by a 100% positive predictive value and 99% negative predictive value.

EyeCatch() in contrast, relying on spatial features alone, exhibited greater stability in its ability to identify eye blink related ICA components despite decreases in the ratio between the artifact amplitude and the noise. However, EyeCatch() exhibited the highest false discovery rate of any of the methods, particularly when the dataset exhibited very low levels of noise, suggesting that under fully automated implementations EyeCatch() might encourage the removal of signal-related ICA components – as reflected by 76.8% positive predictive value and 99.7% negative predictive value.

ADJUST(), which relies on combined stereotypical spatial and temporal features, was observed to exhibit more random failures in the ability to identify ICA components associated with the eye blink, whereas only the likelihood of falsely identifying signal-related ICA components was related to the ratio between the artifact amplitude and the noise. Thus, similar to EyeCatch(), ADJUST() exhibited a 78.2% positive predictive value and 99.9%% negative predictive value suggestive of a bias towards detecting the eye blink related component at the expense of occasionally falsely identifying a signal-related component as artifactual. From a signal detection standpoint these results are sensible: that is, the approach [icablinkmetrics()] that made no false alarms also exhibited many misses, while the approaches [ADJUST() & EyeCatch()] that had the most hits also had the most false alarms.

To ensure that the eye blink artifact is fully removed (e.g., in cases where the ICA algorithm separated the eye blink artifact across multiple components), one might consider the bias to remove several ICA components a strength of the ADJUST() and EyeCatch() approaches. However, within the context of the present investigation, the ICA algorithm was effectively able to dissociate the eye blink related activity into a singular component. Thus, other components simply reflect random perturbations of the signal and their removal would have little benefit for restoring the data to its original uncontaminated state. Indeed, when all three automated approaches returned component identifications, removal of additional components by the ADJUST() and EyeCatch() approaches provided no incremental improvement in restoring the data to its uncontaminated state as all approaches exhibited approximately 91% similarity to the original data following removal of the identified components. Such false positive component identifications, however, may be more detrimental within real EEG datasets as the components selected for removal may be associated with important aspects of the neural signal rather than the artifact. Although the use of simulated data allows for determination of the extent to which these selection approaches can identify the truly artifactual component associated with the eye blink, prior to recommending the utilization of any of these fully automated approaches, it is necessary to further examine their efficacy when used with real EEG data varying across common electrode densities (i.e., 32, 64, and 128 channel montages) and in response to different tasks. We address this issue next.

Generalizability Across Electrode Densities Using Real EEG Data

Method

All participants provided written informed consent in accordance with the Institutional Review Boards at Michigan State University and Binghamton University. The 32 channel dataset included 40 participants (28 female; mean age = 19.6 ± 2.4 years) who performed a go/no go task with images of animals as targets while EEG was recorded (Laszlo & Sacchi, 2015). EEG was digitized at 500 Hz with an A/D resolution of 16 bits and a software filter with a 10 s time constant and a 250 Hz low pass filter with a Brain Amp DC amplifier and a geodesically arranged electro- cap referenced online to the left mastoid and re-referenced offline to averaged mastoids. The VEOG was recorded using an electrode placed on the sub-orbital ridge of the left eye and referenced to the left mastoid.

The 64 channel dataset utilized a sample of 38 participants (20 female; mean age = 19.4 ± 0.9 years) who performed a perceptually challenging three-stimulus oddball task (Pontifex, Parks, Henning, & Kamijo, 2015) while EEG activity was recorded. Continuous data were digitized at a sampling rate of 1000 Hz and amplified 500 times with a DC to 70 Hz filter using a Neuroscan SynampsRT amplifier and a Neuroscan Quik-cap (Compumedics, Inc, Charlotte, NC) referenced to the CCPz electrode of the international 10/5 system (Jurcak, Tsuzuki, & Dan, 2007). Data were re-referenced offline to averaged mastoids. EOG activity was recorded from electrodes placed in a bipolar recording configuration above and below the orbit of the left eye.

The 128 channel dataset utilized a sample of 14 participants (7 female; mean age = 19.2 ± 1.3 years) who performed a free viewing task involving 12 Hz ON/OFF luminance modulated photographs depicting complex, natural scenes while EEG activity was recorded. Continuous data were digitized at a sampling rate of 1000 Hz with 24-bit A/D resolution using the 400 Series Electrical Geodesics, Inc. amplifier (DC to 100 Hz hardware filters) and a 129 HydroCel EEG net referenced to the Cz electrode. EOG activity was recorded from electrodes placed above the orbit of both eyes referenced to the Cz electrode.

These datasets, then, reflect not only diversity in how many electrodes were used, but also in what tasks were performed, how the VEOG was measured, and what configuration montage was used to place electrodes on the scalp. The sample datasets are thus well-suited to addressing how generalizable the results of comparisons between the automated metrics might be to other real-world datasets.

Procedure

For each dataset, the EEG recordings were imported into EEGLAB and prepared for ICA decomposition. Data falling more than two seconds prior to the first event marker and two seconds after the final event marker were removed to restrict computation of ICA components to task related activity. The data were then filtered using a 0.05 Hz high pass IIR filter to remove slow drifts (Mognon, Jovicich, Bruzzone, & Buiatti, 2011). For the 32 and 64 channel datasets, EOG and mastoid (referential) electrodes were removed from the data and relocated in the EEGLAB EEG structure using movechannels() allowing for these electrodes to be restored following removal of the ICA artifact component(s) and the EOG electrodes to be available for use with the icablinkmetrics() function.

ICA decompositions were performed using the extended Infomax algorithm to extract subgaussian components using the default settings called in the binary instance of this function in EEGLAB. Following the ICA computation, each of the three automated procedures (icablinkmetrics() version 3.1, ADJUST() version 1.1.1, and EyeCatch()) were then tested using their default parameters. The artifact channel for icablinkmetrics() was the VEOG channel in the 32 and 64 channel data sets, and channel 25 in the 128 channel EGI system. The icablinkmetrics() function identified eye blinks within the input artifact channel by cross-correlating a canonical eye blink waveform using the eyeblinklatencies() function, only accepting eye blinks which exhibited correlations of 0.96 or higher. As the utilization of real data precludes knowing which components are truly artifactual, components selected by the automated procedures were compared with those selected visually by an expert observer [S.L.] with 12 years of electrophysiology experience who was blind to the selections made by any of the automated approaches. The expert observer followed current standard practice as described within EEGLAB documentation (Delorme & Makeig, 2004), which relies upon visual inspection of the scalp projection maps of the ICA components to make component selections. Thus, while the expert observer [S.L.] was involved in the creation of the icablinkmetrics() algorithm1, in this manner the expert observer’s approach was most similar to the EyeCatch() algorithm. To ensure the integrity of the methodology, if the expert observer and automated procedures disagreed in their classification of components, a more thorough evaluation by an impartial third experienced electrophysiologist (V.M, 10 years experience, who was not involved in the creation of any of the automated selection approaches) was conducted by considering the input from all sources and re-inspecting the data to determine which (if any) were correct. This validation approach is similar to the approaches utilized when validating the ADJUST(), CORRMAP(), and EyeCatch() plugins (Bigdely-Shamlo, Kreutz-Delgado, Kothe, & Makeig, 2013; Mognon, Jovicich, Bruzzone, & Buiatti, 2011; Viola, Thorne, Edmonds, Schneider, & Eichele, 2009). Quantification of the efficacy of the automated component selection approaches for reducing the eye blink artifact were performed by computing the percent reduction in the convolution (i.e., overlap) between the mean eye blink artifact in the EEG collapsed across all electrodes and the EEG activity collapsed across all electrodes during the same period following removal of the selected ICA components.

Statistical Analysis

The efficacy of the automated procedures for identifying the eye blink ICA component were examined statistically by evaluating their sensitivity (the likelihood of correctly identifying the eye blink ICA component(s); i.e., hits) and specificity (the likelihood of correctly not identifying a non-blink component as an eye blink ICA component(s); i.e., correct rejections) relative to the expert-selected component or components. As all datasets were contaminated by eye blink artifact, failure to select an eye blink component was considered a false negative error (‘miss’).

3.2. Results

The mean number of eye blink artifacts present within the data for each participant was 170.5 ± 100.4 (min: 32; max: 490) for the 32 channel data, 107.9 ± 71.6 (min: 30; max: 326) for the 64 channel data, and 61.1 ± 39.9 (min: 20; max: 140) for the 128 channel data. Computation of the independent components was performed using 838.3 ± 251.8 data points for each ICA weight (data points/channels2) for the 32 channel data, 152.9 ± 13.4 points for the 64 channel data, and 35.4 ± 3.0 points for the 128 channel data. The mean time necessary for eye blink identification and metric computation using icablinkmetrics() was 1.0 ± 0.3 seconds for each participant for the 32 channel data, 3.3 ± 1.2 seconds for the 64 channel data, and 7.4 ± 4.0 seconds for the 128 channel data. The mean time necessary for identification of components by the ADJUST() and EyeCatch() algorithms was 2.2 ± 0.6 and 3.1 ± 0.1 seconds for each participant for the 32 channel data, 4.1 ± 0.2 and 6.3 ± 2.4 seconds for the 64 channel data, and 8.2 ± 0.4 and 14.5 ± 0.8 seconds for the 128 channel data, respectively, suggesting that icablinkmetrics() is a slightly faster procedure overall.

Component selection counts along with the sensitivity and specificity are provided in Table 1. When utilizing real EEG data, all three automated procedures exhibited high levels of sensitivity in correctly identifying the eye blink ICA component(s), with both icablinkmetrics() and EyeCatch() exhibiting perfect sensitivity. ADJUST() in contrast, exhibited a sensitivity of 96.7%; failing to identify 1 eye blink component in the 32 channel dataset and 2 components in the 128 channel dataset.

Although perfect sensitivity was observed for both icablinkmetrics() and EyeCatch(), only icablinkmetrics() also exhibited perfect specificity (the likelihood of correctly not identifying a non-blink component as an eye blink ICA component(s); i.e., correct rejections). EyeCatch() falsely identified 89 components (48 from the 32 channel dataset, 21 from the 64 channel dataset, and 20 from the 128 channel dataset) resulting in a false discovery rate of 49.2%. By comparison, ADJUST() falsely identified 198 components (39 from the 32 channel dataset, 123 from the 64 channel dataset, and 36 from the 128 channel dataset) resulting in a false discovery rate of 69%.

To gauge the extent to which removal of the ICA components selected as artifactual was effective in removing the eye blink artifact from the EEG data, the percent reduction in the convolution (i.e., overlap) between the mean eye blink artifact and the EEG activity across all electrode sites during this same period following removal of the selected ICA components was computed. The components selected by icablinkmetrics() and EyeCatch() were observed to reduce the eye blink artifact present within the EEG by 88% and 87.2%, respectively. While the components selected by ADJUST() were observed to reduce the eye blink artifact by 86.6%

Discussion

The aim of this section was to evaluate the generalizability of these automated approaches across real EEG data sets. To this end, we utilized real EEG data recorded with variable numbers of sensors and in response to different experimental tasks. Further, each of the bioamplification systems from which data were submitted utilized a different recording configuration for EOG electrodes (bipolar, lower-orbit unipolar to mastoid, upper-orbit unipolar to vertex) and different acquisition parameters (e.g., acquisition filters, sampling rate, A/D resolution). Despite the substantial diversity in the data provided, icablinkmetrics() demonstrated a high level of performance in automatically identifying the blink-related ICA components, exhibiting perfect sensitivity and specificity. Thus, while icablinkmetrics() exhibited reduced sensitivity under noisy conditions in the simulated data, this noise level would seem to be above that which is normally encountered in real EEG data. The fact that icablinkmetrics() was able to accurately identify eye blink components regardless of the hardware used for data acquisition, EOG montage, EEG montage, or the task being performed by participants, demonstrates its robustness and suggests that it may be suitable for use across a diverse set of data acquisition systems and across tasks.

EyeCatch() similarly exhibited perfect sensitivity in detecting the eye blink related components across datasets. However, when using real EEG data, EyeCatch() rejected a total of 89 non-blink related ICA components (see Figure 5). Some caution is warranted in evaluating this outcome as the EyeCatch() function does not presently offer the ability to differentiate identified eye blink components from lateral eye movement components. Thus, it may be that some of these rejected components reflect truly artifactual, lateral eye movement components which were identified by EyeCatch() but were not the focus of the present investigation. EyeCatch()’s performance on the real data is consistent with its performance on the simulated data, in that its high hit rate was accompanied by a relatively high false alarm rate.

Figure 5.

Figure 5

Two ICA components from a single participant recorded from a 32 channel montage and a single participant recorded from a 128 channel montage. EyeCatch() identified all components as being related to the eye blink, whereas the components on the left of each montage were identified by the Expert Observer and the components on the right of each montage were identified by icablinkmetrics(). Note that for both the example files, the ICA weights are frontally distributed for both components, but only a single component reduces the eye blink artifact when the component is removed. As removal of the additional component has no influence over the mean blink-related activity it can be considered as a false positive component identification.

Despite the popularity of the ADJUST() automated selection routine, the function demonstrated performance below that of either the icablinkmetrics() or EyeCatch(). With ADJUST() rejecting a total of 198 signal related ICA components while still retaining 3 artifact related components, reflected by poorer reduction in the eye blink artifact present within the EEG. Thus, it would appear that the criteria for identifying eye blink related ICA components utilized by the ADJUST() function are neither as specific nor as sensitive as those used by the other automated component selection approaches.

At this point, we have compared the automated methods to each other, in both real and simulated data. However, we have not yet compared them to the commonly used practice wherein trained human observers visually select ICA components. Regardless of the inter-algorithm comparisons, it would not be reasonable to recommend any of them if they cannot out-perform a human observer. For this reason, we next evaluate the performance of trained observers with varying expertise levels in identifying artifactual ICA components in the real data sets used for algorithm comparison in this section.

Accuracy of Component Selection Relative to Trained Observers

Method

The same real EEG datasets used above were used herein to enable comparisons between automated approaches and trained observers. The consensus selected components identified above were compared here with those selected visually by electrophysiologists of varying experience (Expert Observer: 12 years [S.L.]; Competent Observer: 3 years; and Novice Observer: 2 years) who were blind to the selections made by any of the automated approaches. Trained observers had access to the complete EEG dataset in EEGLAB to make their eye blink component selections. Quantification of the efficacy of the trained observers for reducing the eye blink artifact was performed by computing the percent reduction in the convolution (i.e., overlap) between the mean eye blink artifact in the EEG collapsed across all electrodes and the EEG activity collapsed across all electrodes during the same period following removal of the selected ICA components.

Statistical Analysis

The efficacy of the trained observers in identifying the eye blink ICA component were examined statistically by evaluating their sensitivity (the likelihood of correctly identifying the eye blink ICA component(s); i.e., hits) and specificity (the likelihood of correctly not identifying a non-blink component as an eye blink ICA component(s); i.e., correct rejections) relative to the consensus expert-selected component identified above (these were not necessarily always S.L.’s selection, as she could have been over-ruled by consensus between V.M. and the automated approaches). As all simulated datasets were contaminated by eye blink artifact, failure to select an eye blink component was considered a false negative error (‘miss’).

Results

Component selection counts along with the sensitivity and specificity are provided in Table 1. The percent agreement and Fleiss’ Kappa for the selection of the eye blink component amongst the human raters was 95% agreement and 0.316 kappa for the 32 channel dataset, 97.4% agreement and 0.49 kappa for the 64 channel dataset, and 35.7% agreement with 0.125 kappa for the 128 channel dataset.

The expert observer exhibited 96.7% sensitivity, failing to identify 2 eye blink components in the 32 channel dataset and 1 component in the 128 channel dataset. Both the competent observer (with 3 years experience) and the novice observer (with 2 years experience) exhibited 88% sensitivity. The competent observer failed to identify 2 eye blink components in the 32 channel dataset, and 9 components in the 128 channel dataset, while the novice observer failed to identify 2 eye blink components in the 32 channel dataset, 1 component in the 64 channel array, and 8 components in the 128 channel dataset.

Regarding the specificity of the trained observers, even the expert observer incorrectly identified signal related components as being related to the eye blink, exhibiting 99.9% specificity (falsely identifying 6 components: 4 from the 32 channel dataset, and 2 from the 128 channel dataset) with a 6.3% false discovery rate. The competent observer exhibited 99.7% specificity (falsely identifying 15 components: 4 from the 32 channel dataset, 1 from the 64 channel dataset, and 10 from the 128 channel dataset) with a 15.6% false discovery rate. The novice observer exhibited 99% specificity (falsely identifying 48 components: 21 from the 32 channel dataset, 3 from the 64 channel dataset, and 24 from the 128 channel dataset) with a 37.2% false discovery rate.

Discussion

The aim of this section was to assess the accuracy of the commonly used method of trained observers visually selecting ICA components for comparison against the automated methods of eye blink component selection. Across real EEG data varying in number of channels recorded and tasks performed, a clear trend was observed demonstrating the experience-dependent nature of visual component selection. The expert observer was able to correctly identify eye blink components at 96.7% accuracy and rule out non blink related components at 99.9% accuracy. Inexperience was related to decreased accuracy, with 88% accuracy in identifying the eye blink component for both the competent and novice observers, and 99.7% and 99% accuracy in ruling out non blink related components, for the competent and novice observers, respectfully. This experience dependent trend was most readily identifiable within the 128 channel dataset with the number of false discovery component identifications increasing with inexperience, whereas in the 32 and 64 channel datasets there was greater similarity between the expert and competent observers.

Although speculative, the experience dependent nature observed within the 128 channel dataset may be related to the scalp projection maps of the ICA components. A key differentiation between visual inspection of the topographic distribution of ICA weights for 32 and 64 channel data relative to 128 channel data is that by default EEGLAB does not plot the scalp projection maps the same way for the 128 channel plots relative to plots for lower-density arrays. Specifically, the plots are created for 128 channel data without electrode locations and with the activity extending beyond the circumference of the top of the head (as illustrated in Figure 5). Less experienced electrophysiologists may rely on the electrode locations and reference points to a greater extent than more experienced electrophysiologists. Although the default settings were used within the present investigation, it should be noted that it is possible to include the electrode locations using the ‘plotrad’ command in EEGLAB’s topoplot() function to potentially mitigate this issue. Another possible — though equally speculative — reason that there was a larger experience effect for the 128 channel data is that, because so many components are identified in the higher-density ICA computation, the actual eye blink components are somewhat overfit. That is, spatial projections do not display as smooth a topography as the blink components in the lower density data. It may be that less experienced observers rely more on a smoother (less nuanced) template of what the blink artifact should look like, and are thus disproportionately distracted by the spatially overfit components produced in the high density data.

Overall Discussion

Collectively, this investigation sought to determine the efficacy of fully automated approaches for selecting eye blink related temporal ICA components with an eye towards understanding the potential utility of such approaches to replace the labor intensive (and potentially biased) process of human observers manually selecting components. To this end, we assessed the relative strengths of automatic eye-blink ICA component selection methods relying on time-series data [icablinkmetrics()] as compared to those relying on combined stereotypical spatial and temporal features [ADJUST(), (Mognon, Jovicich, Bruzzone, & Buiatti, 2011)], or spatial features alone [EyeCatch(), (Bigdely-Shamlo, Kreutz-Delgado, Kothe, & Makeig, 2013)]. Three questions were then addressed, namely 1) How robust are these approaches to variations in the magnitude of the eye blink artifact amid increasing levels of noise in the signal using simulated EEG data? 2) How generalizable are these approaches across variable electrode densities and experimental tasks? 3) How do these approaches compare to the current common method of trained observers visually selecting temporal ICA components?

Relative to the first two questions, our findings suggest that despite the popularity of ADJUST(), its use of combined stereotypical spatial and temporal features resulted in more random failures in the ability to identify temporal ICA components associated with the eye blink, irrespective of the ratio between the artifact amplitude and the noise when tested using simulated EEG data (Figure 4). When utilized with real EEG data, ADJUST() was able to identify eye blink related components across electrode arrays at a similar level to that of the expert observer (96.7%). However, ADJUST() greatly struggled in the specificity of the component selections exhibiting a false discovery rate of 69%. Thus, while ADJUST() appears to be relatively robust and generalizable in its ability to identify eye blink related ICA components, signal related temporal ICA components representing brain electric activity may be mistakenly identified as artifact and rejected via the ADJUST() approach — particularly when high levels of noise are present within the signal.

Relying on spatial features alone, EyeCatch() was found to be sensitive to variation in the magnitude of the artifact amid increasing levels of noise, demonstrating both a decreased ability to identify eye blink related components when the artifact magnitude to noise ratio was low and an increased occurrence of mistakenly identifying signal related components as artifactual under low noise levels. More promisingly, in the real EEG datasets, EyeCatch() exhibited perfect sensitivity in identifying eye blink related components but struggled in mistakenly identifying signal related components as artifactual with a false discovery rate of 49.2%. The reduced specificity of EyeCatch() was particularly prevalent for data recorded with lower-density electrode montages. Thus, EyeCatch() was observed to be not particularly robust or generalizable in avoiding false discovery component identifications within the current investigation.

The time-series approach utilized by icablinkmetrics(), was the least robust to variation in the amplitude of the artifact relative to increasing levels of noise, particularly struggling to identify the eye blink component when the artifact was small relative to the level of background noise (see Figure 2). However, with the real EEG data, (where the signal to noise ratio was greater than the extremes utilized in the simulation data) icablinkmetrics() exhibited perfect sensitivity across electrode montages and was observed to never falsely identify a component as being related to the eye blink regardless of the use of real or simulated EEG. Thus, while icablinkmetrics() was not observed to be robust in its sensitivity in identifying the artifact under increased levels of noise, it appears both robust and generalizable under more normal EEG recording conditions.

Relative to the third question of how these approaches compare to the current standard method of trained observers visually selecting temporal ICA components, all three methods of automatic component selection were able to accurately identify eye blink related ICA components at or above the level of human observers. Although the expert observer only failed to identify 3 blink components in the 93 real EEG data sets (see Figure 5), potentially more problematic was the finding that 6 ICA components were misidentified as being related to the eye blink, meaning that 6 sources of non-artifactual signal would have been removed from the data (6.3% false discovery rate). Thus, even with substantial experience, signal related temporal ICA components representing brain activity may be mistakenly identified as noise and rejected via the manual rejection approach. This trend is further magnified by expertise level with the competent observer exhibiting a 15.6% false discovery rate, and the novice observer exhibiting a 37.2% false discovery rate while still retaining 11 artifact related components. These findings call into question the extent to which the visual component identification approach using topographic projections of temporal ICA components should be considered the standard. Since visual component identification methods are time consuming and often considered “low level” tasks, they are often relegated to less experienced users. Thus, given the substantially increased false discovery rates demonstrated by inexperienced raters; our findings highlight the reality that when student observers perform visual component identification, they are very likely also removing signal related temporal ICA components from the data which may have substantial ramifications for the post-processed EEG signal. Such an observation is particularly problematic given that only a small number of articles published in Psychophysiology over the past two years using ICA approaches for artifact correction have directly indicated using automated approaches for consistently determining components as artifactual. In the absence of statements in the method, the assumption must therefore be that the vast majority of published literature utilizing the ICA approach for artifact correction has relied on human methods of component selection which are not only resource intensive and slow but may also reduce the quality and integrity of the post-processed EEG signal. Thus, the growing lack of replicability of findings within psychophysiology may very well relate, in some part, to the reliance on human component selections in the increasing number of investigations utilizing ICA.

Recommendations

As with any signal detection problem, these automated techniques must optimize their ability to identify the eye blink related component with their ability to correctly reject components not associated with the eye blink. The use of simulated EEG data within the present investigation highlights a key difference between these automated component selection approaches in this matter. Both ADJUST() and EyeCatch() appear to be optimized towards identifying the eye blink related component at the cost of the occasionally misidentification of a component — as evidenced by demonstrating poorer specificity than even a novice psychophysiologist. Whereas, icablinkmetrics() is optimized towards correctly rejecting components not associated with the eye blink at the cost of occasionally failing to identify the eye blink related component. However, it should be noted that such limitations were not observed with regard to the real EEG data—which fell within a less extreme range of signal to noise ratios than did the simulated data.

Accordingly, any means to utilize these approaches should thus acknowledge their respective limitations. Within the context of ADJUST() and EyeCatch(), these methods would seem better suited towards narrowing down potential candidate eye blink components prior to human inspection. Given the superior performance of EyeCatch() in detecting the eye blink component within real EEG data, a recommendation for implementation would be to have the human observer make visual component selections from those temporal ICA components which were previously identified by EyeCatch(). Rather than sifting through 30 or more components, this automated approach could be utilized to obtain a short list of candidate eye blink components serving to greatly reduce the potential burden and risk of false positive component identification associated with the trained observer approach.

Given the relative strengths of icablinkmetrics(), which avoided false positive component identifications across the diverse data acquisition scenarios of the real EEG and the noisy simulated EEG data, it would seem that this approach is better suited towards a fully automated, user-independent implementation. As the icablinkmetrics() approach either correctly identifies the eye blink component or fails to identify any component, a recommendation for implementation would thus be to have the human observer only visually select components for those datasets where icablinkmetrics() is unable to determine the eyeblink component. Rather than investing time inspecting all datasets, the human observer could instead focus on datasets which are particularly noisy or in which a bad channel was included in the temporal ICA decomposition resulting in reduced quality of signal separation. The icablinkmetrics() approach is ideally suited for such use as in addition to outputting the identified component(s), it also outputs similarity metrics for the eye blink artifact and each ICA component as well as the percent reduction in the eye blink artifact observed when each ICA component is removed along with a graphical output for each ICA component — regardless of whether a component is identified as being artifactual. Accordingly, these metrics could be integrated alongside — rather than in lieu of — topographic projections of temporal ICA weights for those files in which an automatic solution cannot be resolved, or for those files in which visual identification of a component is especially difficult. Thus, in addition to being used as a means to automate the temporal ICA artifact correction approach, icablinkmetrics() could be incorporated as a means of ensuring a high level of confidence during visual selection of components, or to facilitate training of novice electrophysiologists. As recording of eye blink related activity seems to be falling out of favor with newer EEG acquisition systems, a potential weakness of the icablinkmetrics() approach is specification of an electrode channel in which the eye blink artifact manifests most clearly (e.g., the VEOG electrode) in order to construct a template of the eye blink waveform for comparison with the ICA activations. However, it should be noted that any electrode could be used so long as the electrode specified captures the artifact of interest – in the absence of a VEOG electrode per se, several frontal or temporal electrodes would seem to be reasonable alternatives.

Although the present investigation did not assess the efficacy of these automated approaches for handling non-blink related artifacts, it is worth nothing that roughly half of the published literature in Psychophysiology over the past two years indicates only correcting for eye-blink related artifact when using either regression or ICA-based approaches. Ultimately, however, the context of the EEG recording necessarily dictates the nature and degree of artifact present within the data as some protocols may differentially manifest eye blinks, saccadic eye movements, muscle and cardiac related artifacts. While it is clear that investigators are increasingly turning to ICA-based approaches for artifact correction, it is important to emphasize that there is no single best method for correcting all artifacts (Urigüen & Garcia-Zapirain, 2015). ICA-based approaches have been found to be particularly effective for correcting eye blink related artifacts (Jung, et al., 2000). By comparison, regression-based approaches may be more appropriate for other artifacts such as saccadic eye movements and non-stationary artifacts (Hoffmann & Falkenstein, 2008). Based on the current state of the art, an ideal approach may be to implement artifact correction/suppression procedures across multiple processing stages, thereby enabling an investigator to use the best tool for each specific type of artifact (Urigüen & Garcia-Zapirain, 2015). Indeed, the use of ICA and regression-based approaches for artifact correction are not inherently mutually exclusive. Given the relative strengths and weaknesses of these methods (Hoffmann & Falkenstein, 2008), a temporal ICA approach to eye-blink artifact correction could be combined with existing regression-based approaches for the correction of other non-blink related artifacts.

In considering the automated selection of temporal ICA components, it is important to note that a limitation of the present investigation is that the efficacy of these automated approaches may vary based upon the particular characteristics of the artifact of interest. Artifacts that exhibit temporally consistent morphological characteristics (such as eye blink and ECG artifacts) would seem well suited for correction using temporal approaches to component identification. In such instances, the stationarity of the artifact produces a cleaner isolation and is ideally suited for time-domain approaches to component selection since the individual artifacts temporally align with the artifactual component(s). However, other non-stationary artifacts (such as saccadic eye movements) may be better suited for spatial approaches to component identification or for regression-based artifact correction procedures (Hoffmann & Falkenstein, 2008).

Conclusions

Collectively, the present investigation demonstrated the efficacy of utilizing automated approaches for temporal ICA eye blink artifact selection, and compared automated approaches directly with human selection of eye blink components amongst psychophysiologists with a range of expertise. All of the automated methods assessed were good enough at identifying artifactual components to be considered as candidates for supplementing or replacing manual inspection. However, icablinkmetrics(), in particular, would seem to provide an effective means of automating eye blink correction using temporal ICA while at the same time eliminating human errors inevitable during manual component selection and false positive component identifications common in other automated approaches, given its exceptional specificity in all cases.

Supplementary Material

Supp Fig S1
Supp Fig S2
Supp info

Acknowledgments

Support for the preparation of this manuscript was provided by a grant from the Eunice Kennedy Shriver National Institute of Child Health and Human Development (NICHD) to Matthew Pontifex (R21 HD078566), and by grants from the National Science Foundation (NSF) to Sarah Laszlo (NSF CAREER-1252975, NSF TWC SBE-1422417, and NSF TWC SBE-1564046).

Footnotes

1

Though S.L. contributed to the design of icablinkmetrics(), she was not responsible for actually implementing it and did not know what its behavior would be with respect to these datasets prior to making her selections.

Author Disclosure Statement

No conflicting financial interests exist.

References

  1. Bell AJ, Sejnowski TJ. An information maximisation approach to blind separation and blind deconvolution. Neural Computation. 1995;7:1129–1159. doi: 10.1162/neco.1995.7.6.1129. [DOI] [PubMed] [Google Scholar]
  2. Bigdely-Shamlo N, Kreutz-Delgado K, Kothe C, Makeig S. EyeCatch: Data-mining over half a million EEG independent components to construct a fully-automated eye-component detector. Annual International Conference of the IEEE Engineering in Medicine and Biology Society; Piscataway, NJ. IEEE Service Center; 2013. pp. 5845–5848. [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Castellanos NP, Makarov VA. Recovering EEG brain signals: Artifact supression with wavelet enhanced independent component analysis. Journal of Neuroscience Methods. 2006;158:300–312. doi: 10.1016/j.jneumeth.2006.05.033. [DOI] [PubMed] [Google Scholar]
  4. Delorme A, Makeig S. EEGLAB: an open source toolbox for analysis of single-trial EEG dynamics. Journal of Neuroscience Methods. 2004;134:9–21. doi: 10.1016/j.jneumeth.2003.10.009. [DOI] [PubMed] [Google Scholar]
  5. Flexer A, Bauer H, Pripfl J, Dorffner G. Using ICA for removal of ocular artifacts in EEG recorded from blind subjects. Neural Networks. 2005;18:998–1005. doi: 10.1016/j.neunet.2005.03.012. [DOI] [PubMed] [Google Scholar]
  6. Hoffmann S, Falkenstein M. Correction of eye blink artefacts in the EEG: A comparison of two prominent methods. PLOS One. 2008;3:1–11. doi: 10.1371/journal.pone.0003004. [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Joyce CA, Gorodnitsky IF, Kutas M. Automatic removal of eye movement and blink artifacts from EEG data using blind component separation. Psychophysiology. 2004;41:313–325. doi: 10.1111/j.1469-8986.2003.00141.x. [DOI] [PubMed] [Google Scholar]
  8. Jung T, Makeig S, Humphries C, Lee T, McKeown MJ, Iragui V, Sejnowski TJ. Removing electroencephalographic artifacts by blind source separation. Psychophysiology. 2000;37:163–178. [PubMed] [Google Scholar]
  9. Jurcak V, Tsuzuki D, Dan I. 10/20, 10/10, and 10/5 systems revisited: Their validity as relative head-surface-based positioning systems. NeuroImage. 2007;34:1600–1611. doi: 10.1016/j.neuroimage.2006.09.024. [DOI] [PubMed] [Google Scholar]
  10. Laszlo S, Armstrong BC. PSPs and ERPs: Applying the dynamics of post-synaptic potentials to individual units in simulation of ERP reading data. Brain and Language. 2014;132:22–27. doi: 10.1016/j.bandl.2014.03.002. [DOI] [PubMed] [Google Scholar]
  11. Laszlo S, Plaut DC. A neurally plausible parallel distributed processing model of event-related potential word reading data. Brain and Language. 2012;120:271–281. doi: 10.1016/j.bandl.2011.09.001. [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Laszlo S, Sacchi Individual differences in involvement of the visual object recognition system during visual word recognition. Brain and Language. 2015;145:42–52. doi: 10.1016/j.bandl.2015.03.009. [DOI] [PubMed] [Google Scholar]
  13. Luck SJ. An introduction to the event-related potential technique. MIT press; 2014. [Google Scholar]
  14. Mognon A, Jovicich J, Bruzzone L, Buiatti M. ADJUST: An automatic EEG artifact detector based on the joint use of spatial and temporal features. Psychophysiology. 2011;48:229–240. doi: 10.1111/j.1469-8986.2010.01061.x. [DOI] [PubMed] [Google Scholar]
  15. Pontifex MB, Parks AC, Henning DA, Kamijo K. Single bouts of exercise selectively sustain attentional processes. Psychophysiology. 2015 doi: 10.1111/psyp.12395. [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Urigüen JA, Garcia-Zapirain B. EEG artifact removal—state-of-the-art and guidelines. Journal of Neural Engineering. 2015;12:1–23. doi: 10.1088/1741-2560/12/3/031001. [DOI] [PubMed] [Google Scholar]
  17. van Drongelen W. Signal processing for neuroscientists: An introduction to the analysis of physiological signals. Burlington, M.A.: Elsevier; 2007. [Google Scholar]
  18. Viola FC, Thorne J, Edmonds B, Schneider T, Eichele T. Semi-automatic identification of independent components representing EEG artifact. Clinical Neurophysiology. 2009;120:868–877. doi: 10.1016/j.clinph.2009.01.015. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supp Fig S1
Supp Fig S2
Supp info

RESOURCES