Abstract
Over 15 million epilepsy patients worldwide do not respond to drugs. Successful surgical treatment requires complete removal, or disconnection of the seizure onset zone (SOZ), brain region(s) where seizures originate. Unfortunately, surgical success rates vary between 30%-70% because no clinically validated biological marker of the SOZ exists. We develop and retrospectively validate a new EEG marker - neural fragility - in a retrospective analysis of 91 patients by using neural fragility of the annotated SOZ as a metric to predict surgical outcomes. Fragility predicts 43/47 surgical failures with an overall prediction accuracy of 76%, compared to the accuracy of clinicians being 48% (successful outcomes). In failed outcomes, we identify fragile regions that were untreated. When compared to 20 EEG features proposed as SOZ markers, fragility outperformed in predictive power and interpretability suggesting neural fragility as an EEG biomarker of the SOZ.
1. Introduction
Over 15 million epilepsy patients worldwide and 1 million in the US suffer from drug-resistant epilepsy (DRE) [1, 2]. DRE is defined as continued seizures despite two trials of appropriately chosen anti-epileptic drugs [3]. DRE patients have an increased risk of sudden death and are frequently hospitalized, burdened by epilepsy-related disabilities, and their cost of care is a significant contributor to the $16 billion dollars spent annually in the US treating epilepsy patients [4]. Approximately 50% of DRE patients have focal DRE, where specific brain region(s), termed the epileptogenic zone (EZ), is necessary and sufficient for initiating seizures and whose removal (or disconnection) results in complete abolition of seizures [5, 6]. The EZ encompasses the clinically identified seizure onset zone (SOZ) and early propagation zone (EPZ). The brain regions associated with the SOZ demonstrate the earliest electrophysiological changes during a seizure event, and in general precede the clinical onset of seizures; and the EPZ regions are involved at the time of the earliest clinical (semiological) manifestations during a seizure event. Successful surgical and neuromodulatory treatments can stop seizures altogether or allow them to be controlled with medications [7], but outcomes for both treatments critically depend on accurate localization of the SOZ.
Localizing the SOZ also relies on accurate placement of electrodes such that they cover the EZ, and the ability to identify abnormalities in the intracranial EEG (iEEG) channels that may correlate to the SOZ with the naked eye. Unfortunately, even the most experienced clinicians are challenged because epilepsy is fundamentally a network disease, which cannot be entirely defined by the current methods of localization. Abnormal connections across several channels may constitute a more effective marker of the SOZ [8]. Localization thus lends itself to a data-driven network-based computational approach and several electroencephalogram (EEG) algorithms have been proposed to localize the SOZ from recordings. Many entail investigations of the spectral power in each iEEG channel including high frequency oscillations [9], but these approaches do not consider network properties of the brain because they treat each EEG channel independently. Others have proposed graph-based analysis of iEEG [10, 11, 12, 13, 14, 15], but these approaches fail to identify internal network properties that cause seizures to occur in the first place.
We propose an EEG marker of the SOZ, neural fragility (conceptually described in Figure 1 and quantitatively in Figure 2). To create the fragility marker, we first build a personalized dynamic model of the brain network from observed iEEG signals (top row Figure 2). The generative model can accurately reconstruct the patient’s iEEG recordings [16, 17]. We then calculate neural fragility which measures the degree network nodes are imbalanced, i.e., small impulse perturbations on the network and thus can trigger seizures (see Figure 1) [18]. To evaluate neural fragility as a marker for the SOZ, we conduct a retrospective study using iEEG data from 91 patients treated across 5 epilepsy centers: Johns Hopkins Hospital (JHH), National Institute of Health (NIH), Cleveland Clinic (CClinic), University of Maryland Medical Center (UMMC) and Jackson Memorial Hospital of University of Miami (UMF). In the study population, all DRE patients underwent invasive iEEG monitoring followed by surgical resection or laser ablation of the SOZ (44 success and 47 failure outcome). We demonstrate that neural fragility is higher/lower in electrode contacts within clinically annotated SOZs for success/failure patients. In addition, we compare fragility of iEEG nodes to 6 frequency-based and 14 graph theoretic features in a 10-fold nested-cross validation experiment. Neural fragility has an area under the curve (AUC) discrimination score of 0.88 +/− 0.064, which is 13% better compared to the next best feature. In addition, it has a high degree of interpretability, which we demonstrate by computing an interpretability ratio suggesting that spatiotemporal heatmaps of neural fragility can be a robust iEEG biomarker of the SOZ, seamlessly integrated into the clinical workflow.
2. Results
There is no clinically validated biomarker of the SOZ, yet it is an important component for the localization of the underlying EZ [19]. This presents a serious challenge for clinicians to accurately localize the SOZ and has led to surgical success rates to vary between 30-70% despite large brain regions being removed [20]. With no biomarker available, clinicians conduct extensive evaluations with neuroimaging, clinical testing, and visual inspection of EEG recordings (Figure 3). When non-invasive tests are inconclusive, patients undergo intracranial monitoring, when iEEG electrodes are implanted into the brain. iEEG provides high temporal resolution data for clinicians to visually detect abnormal activity, such as spikes and high frequency bursts, in between seizures (interictal) and during seizures (ictal) (see an outline in Supplementary Figure 1 and Supplementary Figure 2).Specifically, clinicians attempt to identify electrodes involved in the SOZ and early spread [21]. Unless the region overlaps with eloquent cortex, surgical resection is then performed on the basis of this hypothesis.
We analyze every patient’s iEEG using fragility and 20 other baseline features, resulting in spatiotemporal heatmaps for every feature. The baseline features include spectral power in various frequency bands (e.g. delta band 1-4 Hz) and specific graph measures of bivariate correlation measures (e.g. eigenvector centrality and degree of correlation and coherence matrices), which have been previously reported in the literature to correlate to the SOZ [11, 12, 13, 10]. We evaluate each feature by predicting surgical outcome (process outlined in Extended Data Figure 1).
2.1. Description of Neural Fragility
Neural fragility is a paradigm shift in the EEG analytics space. It is a concept based on the conjecture that focal seizures arise from a few fragile nodes, i.e., the SOZ, which renders the cortical epileptic network on the brink of instability. When one observes iEEG data during interictal, or preictal periods, activity recorded from each channel appears to hover around a baseline value (Figure 1 left). If the network is “balanced”, then it will respond transiently to an impulse, but always returns to a baseline value (e.g. Figure 2ab). In contrast, when one observes iEEG data during a seizure event, activity (i) grows in amplitude, (ii) oscillates, and (iii) spreads in the brain when the network is perturbed (Figure 1 right). This is a consequence of an “unbalanced” network that does not return to a baseline value (e.g. Figure 2c). From a dynamical systems perspective, the iEEG network has switched from a stable (non-seizure) to an unstable (seizure) network.
Biologically, imbalance due to perturbations between excitatory and inhibitory connections of a neural network can occur through any number of mechanisms, such as elevated glutamate [22, 23], genetic disorder impacting synaptic inhibition [24], decreased GABA [25], inclusion of axo-axonic gap junctions [26], loss of inhibitory chandelier cells [27], or axonal sprouting from layer V excitatory pyramidal cells [28]. This imbalance within a neural network may lead to functional instability, where impulses at certain nodes lead to recurring seizures. While iEEG cannot distinguish between excitatory and inhibitory neuronal populations, the concept of imbalance causing the network to be on the brink of instability can be modeled by neural fragility at the iEEG network level.
To demonstrate how fragility is computed from a dynamical model, we consider a 2-node network as shown in Figure 2. In Figure 2a), a stable network is shown where excitation and inhibition are balanced. The network model is provided in the top row and takes a linear form, x(t + 1) = Ax(t), where t is a time index (typically one millisecond). When the inhibitory node is stimulated by an impulse, both nodes transiently respond and the EEG returns to baseline (bottom row in Figure 2). In panel b), the inhibitory node’s connections are slightly perturbed in a direction that makes the inhibitory node less inhibitory (see changes to its connectivity to the excitatory node). Now, when the inhibitory node is stimulated by an impulse, the responses from each node have a larger transient response but still return to baseline. Finally, in panel c), the inhibitory node’s connections are further perturbed in a direction that makes the inhibitory node less inhibitory. Now, when the inhibitory node is stimulated by an impulse, the responses oscillate demonstrating that the network has gone unstable. The fragility of the inhibitory node is thus quantified as which is the norm of the perturbation vector applied to the first column of A. Our conjecture is that small changes in connection strengths at SOZ nodes cause an imbalance in connectivity between nodes in the network, resulting in susceptible seizures.
To test our conjecture, we estimate a linear time-varying model with a sliding window of A matrices, that characterize a linear dynamical system: x(t + 1) = Ax(t) [17, 16]. This is a generative model representing the linear dynamics between iEEG channels within a small window of time. Each A matrix is estimated from data via a least-squares method. We have shown that this linear approximation is a valid model of the data [17]. From this model, we compute the minimum 2-norm perturbation matrix, , over all N channels [16] (see Extended Data Figure 1). The 2-norm of each perturbation matrix represent the neural fragility for that channel. Computing the fragility now across sequentially estimated A matrices results in a spatiotemporal neural fragility heatmap. For a full description of neural fragility, see Methods - section Neural fragility of iEEG network.
2.2. Fragility highlights the clinical SOZ in successful patients
To qualitatively assess the usefulness of fragility in localizing electrodes of interest, we first look at specific examples of patients with varying outcomes and epilepsy etiologies (see Figure 3). We analyzed with fragility and demonstrate how it may provide additional information for SOZ localization. In Figure 4, we show three patients with differing surgical treatments, outcomes, Engel class and clinical complexity along with their fragility heatmaps and corresponding raw iEEG data (for full clinical definitions; see Supplementary Excel table).
In Figure 4a, the red electrode labels on the y-axis correspond to the clinical SOZ electrodes; the red electrodes are typically a subset of the resected region. It shows the period 10 seconds before and after electrographic seizure onset (black dashed line). In Patient_1, the fragility heatmap agreed with clinical visual EEG analysis, identifying the SOZ, which was surgically resected and led to a seizure free patient. The clinical SOZ has high fragility, even before seizure onset, which is not visible in the raw EEG. The heatmap for Patient_1 also captures the propagation of seizure activity (Extended Data Figure 2). This patient had a successful surgery, and so we can assume the resected tissue likely contained the epileptic regions; it is likely clinicians correctly localized the EZ. When viewing the raw EEG data in Figure 4b (top), Patient_1 has iEEG signatures that are readily visible around seizure EEG onset (Figure 4b). We see high-frequency and synchronized spiking activity at onset that occurs in electrodes that clinicians annotated as SOZ, which correspond to the most fragile electrodes at onset. In addition, the fragility heatmap captures the onset (ATT and AD electrodes) and early spread of the seizure (PD electrodes). Moreover, ATT1 (anterior temporal lobe area) shows high fragility during the entire period before seizure onset (Figure 4). This area was not identified with scalp EEG, or non-invasive neuroimaging.
2.3. Fragility of regions outside the SOZ in failed patients
In Figure 4, Patient_26 and Patient_40 both show regions with high fragility that were not in the clinically annotated SOZ (or the resected region), and both had recurrent seizures after surgery. From seizure onset, many electrodes in both patients exhibit the clinically relevant EEG signatures, such as spiking and fast-wave activity [29]. In the raw EEG, one can see synchronized spikes and spike-waves, but ultimately the epileptic regions were not successfully resected. Visual analysis of the EEG was insufficient for Patient_26 and Patient_40, which ultimately led to insufficient localizations and failed surgical outcomes.
Patient_26 had a resection performed in the right anterior temporal lobe region where clinicians identified the SOZ (RAD, RHD and RTG40/48 electrodes). However, ABT (anterior basal temporal lobe), PBT (posterior basal temporal lobe) and RTG29-39 (mesial temporal lobe) electrodes were highly fragile compared to the clinical SOZ, but not annotated as SOZ. In the raw EEG shown in Figure 4b, it is not apparent that these electrodes would be part of the SOZ. Patient_40 had laser ablation performed on the electrode region associated with Q2, which was not fragile. However, X’ (posterior-cingulate), U’ (posterior-insula) and N’/M’/F’ (superior frontal gyrus) electrodes were all fragile compared to Q2, which recorded from a lesion in the right periventricular nodule. Based on these heatmaps, the fragile regions could be hypothesized to be part of the SOZ, and possibly candidates for resection.
2.4. Fragility outperforms other features in predicting outcomes
To test the validity of neural fragility and the baseline features as SOZ markers, we investigate each feature’s ability to predict surgical outcomes of patients when stratified by the set of SOZ contacts and the rest which we denote as the SOZ complement, SOZC (for a distribution of patient outcomes, see Supplementary Figure 3 and Table 1). Neural fragility in the SOZ was significantly higher then the SOZC in successful patients as compared to failed patients (Success pvalue = 3.326e-70, Fail pvalue = 0.355; Extended Data Figure 3). Patient_01 has higher neural fragility in the SOZ when compared to Patient_26 and Patient_40 (Extended Data Figure 4). This effect is seen when pooling patients across all centers as well, where neural fragility is either i) higher before the seizure onset, or ii) has a marked increase starting at seizure onset (Extended Data Figure 5). To evaluate this at a single subject level, we train a Structured Random Forest model (RF) for each feature that takes in a partitioned spatiotemporal heatmap based on the clinically annotated SOZ and generates a probability of success - a confidence score in the clinical hypothesis (for full details, see Methods Section). We test each feature’s model on a held out data set by applying a threshold to the model’s output and computing a receiver operating characteristic (ROC) curve. The ROC curve plots true positive rates versus false positive rates and the area under the curve (AUC), a measure of predictive power of the feature. We also compute the precision (PR) or positive predictive value (PPV) for each feature, which is the proportion of predicted successful or “positive” results that are true positives (actual successful surgeries). In addition, we compute the negative predictive value (NPV) as well. The larger the AUC/PPV/NPV is for a feature, the more predictive it is and thus more valid as an iEEG marker of the SOZ.
Table 1:
Clinical Center Patients | |||||||
---|---|---|---|---|---|---|---|
| |||||||
Clinical Covariate | NIH | JHH | UMMC | UMH | CClinic | Total | |
| |||||||
Number Patients | 14 | 4 | 7 | 5 | 61 | 91 | |
Number Success | 9 | 1 | 7 | 3 | 24 | 44 | |
Number Failure | 5 | 3 | 0 | 2 | 37 | 47 | |
Number Male | 6 | n/a | 7 | 1 | 30 | 44 | |
Number Female | 8 | n/a | 0 | 4 | 31 | 43 |
Since each patient’s implantation has varying number of electrodes, we summarize each feature’s distribution of the SOZ and SOZC electrodes into quantile statistics over time (see Methods - Experimental design). As an input to the RF models, we compute the quantile statistics over time of the SOZ and SOZC set for each patient and each feature, which are input into a machine learning RF model. RF models are attractive because they are non-parametric, interpretable (they are a set of decision trees performing a consensus procedure), and able to handle higher dimensional data better compared to other models, such as Logistic Regression [30]. As an output of the trained RF model, each heatmap gets mapped to a probability that the outcome will be a success. An RF model is tuned for each feature through 10-fold cross-validation (CV; see Methods - Experimental design), resulting in a uniform benchmark against neural fragility on the same set of patients. Similar to neural fragility, all benchmark features are computed as heatmaps (Supplementary Figure 4).The “high-gamma” frequency band feature encompasses what some would consider HFOs (i.e. 90-300 Hz).
In terms of AUC (measures discrimination), neural fragility performs the best (Figure 5a). Compared to the top 3 leading baseline features, neural fragility outperforms the beta band (15-30 Hz) power by more than 13%. The AUC of fragility is the highest with a value of 0.88 +/− 0.064, compared to the next best representation, the beta frequency band, with a value of 0.82 +/− 0.040. From the ROC and PR curves, we observe that the fragility consistently has a higher sensitivity for the same false positive rate compared to the 20 other feature representations (Extended Data Figure 6). In terms of effect size, neural fragility improves over the beta frequency band with a large effect size of 0.97 Cohen’s D with 9 out of 10 folds improving the AUC (Paired Wilcoxon Rank-sum PValue = 0.027 (Extended Data Figure 6).
In terms of PR (measures PPV), neural fragility also performs the best (Figure 5b). In terms of average precision, which weighs the predictive power in terms of the total number of patients, neural fragility also obtains an average precision of 0.83 +/− 0.076, which is >5% better than the next best feature. In addition, compared to the clinical surgical success rate of 47%, fragility had a 76% +/− 6% accuracy in predicting surgical outcome, a PPV of 0.903 +/− 0.103 and a NPV of 0.872 +/− 0.136. In terms of PR, neural fragility improves over the beta band power with a medium effect size of 0.57 Cohen’s D with 8 out of 10 folds improving the PR (PValue = 0.08).
When we compare the success probabilities predicted by the models between surgical outcomes, we observe that neural fragility separates success and failed outcomes the most with a Cohen’s D of 1.507 (1.233-1.763; 95% confidence interval) (Figure 5c) and pvalue of 6.748e-31 (Figure 5d). In addition to having good discrimination, we compute how well-calibrated the model is - that is having the success probabilities values reflect true risk strata in the patient population. For example, a perfectly calibrated model would have a 20% confidence in a set of patients with exactly 20% success outcomes. We quantify how well calibrated the success probability distributions are over the held-out test set and find that the RF model trained on neural fragility produces well-calibrated predictions (Extended Data Figure 6). The Brier-loss (a measure of calibration; 0 being perfect) was 0.162 +/− 0.036, which was a 6.7% improvement compared to the next best feature.
2.5. Fragility correlates with important clinical covariates
Neural fragility also correlates with more granular treatment outcomes and clinical complexity. Successful seizure-free outcomes (Engel I and ILAE 1) and failure seizure-recurrence outcomes (Engel II-IV) are defined at 12+ months post-op. In addition, we can categorize patients by their clinical complexity (CC) as: (1) lesional, (2) focal temporal, (3) focal extratemporal, and (4) multi-focal (Figure 3a) [21]. We stratify the distribution of the success probabilities based on Engel score and show a decreasing trend in success probability as Engel score increases (Figure 6b). The effect sizes when comparing against Engel I were 1.067 for Engel II (P = 4.438e-50), 1.387 for Engel III (P = 2.148e-69), and 1.800 for Engel IV (P = 4.476e-74). Although the AUC indirectly implies such a difference would exist between Engel I (success) and Engel II-IV (failures), we also observe that confidence decreases as the severity of the failure increases; Engel IV patients experience no changes in their seizures, suggesting the true epileptic tissue was not localized and resected, while Engel II patients experience some changes, suggesting there was portions of the true epileptic tissue that may have been resected. We also compare the success probability distributions with respect to the ILAE score, which is another stratification of the surgical outcomes (Figure 6c). There is a similar trend as seen in the Engel score - as ILAE score increases, the success probability decreases.
We also analyze the success probability with respect to the epilepsy severity measured by the clinical complexity (CC) of the patient, which is a categorization of the etiology of the disease. CC is determined by what type of seizures the patient exhibits rather than the severity of the seizures after-surgery (see Methods - Dataset collection). CC is a factor that is determined before surgery and one we expect could correlate with failure rate. The higher the CC, the more difficult localization is and hence more likely seizures recur after surgery. In Figure 6a, we observe that CC1 and CC2 (i.e. lesional and temporal lobe) patients had a similar distribution of success probabilities (Cohen’s D = 0.020; MW U-test PValue = 0.799; LQRT PValue = 0.76), while CC3 and CC4 had significantly lower distributions (CC3: [Cohen’s D = 0.779; MW U-test PValue = 2.808e-19; LQRT PValue = 0.00]; CC4: [Cohen’s D = 1.138; MW U-test PValue = 7.388e-26; LQRT PValue = 0.00]). This trend is not optimized for directly in our discrimination task, but it aligns with clinical outcomes. CC1 and CC2 are comparable, which agrees with current data suggesting that lesional and temporal lobe epilepsy have the highest rates of surgical success [31, 32]. Extratemporal (CC3) and multi-focal (CC4) patients tend to have lower success rates due to difficult localizations and thus the neural fragility confidence in those SOZ localizations should be low [31].
The highest values of neural fragility (i.e. red zones) that differentiates the SOZ and SOZC contribute the most to making a correct classification (Extended Data Figure 7). In general neural fragility is not correlated with any single frequency band (Extended Data Figure 8). We also show how fragility heatmaps are the most interpretable over all baseline features. When comparing the contrast between SOZ and SOZC of success and failed patients, neural fragility has the largest difference, whereas beta power is hard to distinguish (Extended Data Figure 9). Finally, we examine the model predictions based on gender, handedness, onset age and surgery age to show that there are no relevant differences (Extended Data Figure 10).
3. Discussion
We demonstrate that neural fragility, a networked-dynamical systems based representation of iEEG is a strong candidate for a highly interpretable iEEG biomarker of the SOZ. We compared neural fragility to 20 other popular features using data from n=91 patients treated across five centers with varying clinical complexity. Neural fragility performed the best in terms of AUC, PR and interpretability, when bench-marking proposed iEEG features in a uniform fashion over the same sets of patients using the best possible parameters of a RF model to compute a confidence measure (i.e. probability of success) of the clinically annotated SOZ. We spent over four years to collect and annotate this heterogeneous dataset. To facilitate reproducibility and further investigation of iEEG features, we made this dataset BIDS-compliant and publicly available through the OpenNeuro website (for details see Methods - Code availability).
Challenges in validating iEEG features as SOZ markers
Many features have been proposed as potential biomarkers for the SOZ, but none have successfully translated into the clinical workflow [11, 12, 33, 31]. Current limitations for evaluating computational approaches to localization are largely attributed to i) insufficient benchmarking to other iEEG features and ii) a lack in sufficient sampling across epilepsy etiologies.
Even with seemingly successful features derived from data, it is important to benchmark against existing approaches to provide a holistic view of the value of the said feature. Without benchmarking, it is easy to become overly optimistic in terms of the performance of a feature, whereas it is possible that other iEEG features perform just as well due to the low clinical complexity presentation of the patient. While other traditional features such as the power in the beta band seem to be informative in SOZ localization, neural fragility outperforms in effect size, p-value and interpretability.
Although predicting surgical outcomes in our experimental setup is promising, it will be important to understand why certain localizations are successful and why certain are insufficient. For example, in patients with lesions on MRI scans that correlate with the patient’s EEG and seizure semiology, surgical resection can lead to seizure freedom in approximately 70% of patients [21]. Even in these relatively straightforward cases, localization is not perfect, possibly due to chronic effects of epilepsy such as kindling, which can cause neighboring tissue to become abnormal and epileptogenic [34]. There is a need to sample heterogeneous and large patient populations and derive a feature invariant to epilepsy type and clinical covariates.
Clinical case complexity and surgical outcomes
In our dataset, we saw varying outcomes across the five clinical centers. Difference in seizure outcomes across the clinical centers can be explained by several factors affecting the complexity of the epileptic syndromes. The multi-factorial contributory factors can be related to i) the percentage of non-lesional versus lesional cases, ii) multiple surgical interventions in the past, iii) patient selection and iv) group experience. For example, JHH is a tertiary referral center with a great deal of experience, but also high complexity cases (non-lesional MRIs, multiples surgical interventions in the past, complex semiology and EEG interpretation). In addition, data collection is limited depending on clinician resources and retrospective data availability. The relative center-by-center outcomes are not reflective of the actual center’s surgical outcome rate, but rather samples from the clinical cases that those clinicians analyzed and were readily available for sharing. The pooled 91 patients though do reflect the approximately 50% surgical success rates seen in DRE patients [20, 35, 36]. By including multiple centers, we sought to build a diversified sample of varying clinical complexities and practices, thus lending the evaluation of neural fragility more confidence. An important next step would be the prospective evaluation of neural fragility in the context of different clinical case complexities to determine if surgical treatment can be improved with SOZ localization assistance. In addition, future datasets should be made BIDS-compliant and available to the public to further developments of localization algorithms.
Limitations of the most popular iEEG features
A google scholar search using the keywords: “localization of seizure onset zone epilepsy intracranial EEG” produces over 24,000 results. This is striking considering no computational tools to assist in SOZ localization are in the clinical workflow today. The majority of proposed features lack consistency in finding objective iEEG quantities that correlate to clinically annotated SOZ because they fail to capture internal properties of the iEEG network which are critical to understand when localizing the SOZ. Proposed algorithms either (i) compute EEG features from individual channels (e.g. spectral power in a given frequency band), thus ignoring dependencies between channels [9, 37] to name a few, or they (ii) apply network-based measures to capture pairwise dependencies in the EEG window of interest [11, 12]. Specifically, correlation or coherence between each pair of EEG channels is computed and organized into an adjacency matrix, on which summary statistics are derived including degree distribution and variants of centrality [11, 12, 13, 10, 15]. Such network-based measures are not based on well formulated hypotheses of the role of the epileptic tissue in the iEEG network, and many different networks (adjacency matrices) can have identical summary statistics resulting in ambiguous interpretations of such measures [38].
A popular EEG feature that has been proposed as an iEEG marker of the SOZ and reported in over 1000 published studies is High-frequency oscillations (HFOs) ([9, 39, 33] to name a few). HFOs are spontaneous events occurring on individual EEG channels that distinctively stand out from the background signal and are divided into three categories: ripples (80–250 Hz), fast ripples (250–500 Hz), and very-fast ripples (>500 Hz) [40]. Retrospective studies suggested that resecting brain regions with high rates of HFOs may lead to good post-surgical outcome (e.g., [41, 42]). Although they found significant effects for resected areas that either presented a high number of ripples or fast ripples, effect sizes were small and only a few studies fulfilled their selection criteria [42]. Furthermore, several studies have also questioned the reproducibility and reliability of HFOs as a marker [43, 41, 39]. In addition, there are also physiologic, non-epileptic HFOs, which poses a challenge, as disentangling them from pathological HFOs is still an unsolved issue [44].
Similar inconclusive results hold in completed prospective studies of HFOs. In 2017, an updated Cochrane review [45] investigated the clinical value of HFOs regarding decision making in epilepsy surgery. They identified only two prospective studies at the time and concluded that there is not enough evidence so far to allow for any reliable conclusions regarding the clinical value of HFOs as a marker for the SOZ. Today, five clinical trials are listed as using HFOs for surgical planning on clinicaltrials.gov as either recruiting, enrolling by invitation, or active and not enrolling and none have reported results. The fundamental limitation of the aforementioned studies lies in the fact that they approach the SOZ EEG marker discovery process as a signal processing and pattern recognition problem, concentrating on processing EEG observations to find events of interest (e.g. HFOs) as opposed to understanding how the observations were generated in the first place and how specific internal network properties can trigger seizures.
Why Neural Fragility Performs Well
Rather than analyzing iEEG data at the channel and signal processing level, we seek to model the underlying network dynamics that might give rise to seizures in the form of neural fragility in a dynamical network model. A notion of fragility in networks is commonly seen in analysis of structural [46] network. Although we are not directly analyzing the structural nature of neuronal network, there are studies that have characterized epilepsy in terms of structural fragility and network organization [47, 18]. Specifically, in cellular studies [28, 27], epilepsy is caused by changes in the structural cell network, which causes loss of inhibition or excessive excitation; these biological changes cause downstream aberrant electrical firing (i.e. seizures). By quantifying the fragility of each node in an iEEG network, we determine how much of a change in that region’s dynamical connections is necessary to cause seizure-like phenomena (e.g. instability). As a result, high neural fragility coincides with a region that is sensitive to minute perturbations, causing unstable phenomena in the entire network (i.e. a seizure).
Scientific Advances Emerging from Neural Fragility
Neural fragility has the potential to re-define how epilepsy surgery is performed, departing from the classical “localization paradigms” and “en-bloc resections” to a personalized “network-based” user-friendly visualization and surgical strategy. By developing a 3D (brain region, time, fragility) network-based method for anatomical representation of the epileptiform activity, including the seizure onset areas and the early propagation zone, this study will have the potential to offer a safer, more efficient, and cost-effective treatment option for a highly challenging group of patients with disabling DRE. More precise SOZ localization using neural fragility would also guide of chronic implantation of neurostimulation devices aimed to suppress seizures with bursts of current when detected [48, 49].
Neural fragility may also be relevant in detecting epileptogenic regions in the brain when applied to interictal (between seizures) iEEG recordings. Seizure iEEG data are currently the gold standard in clinical practice for localizing the SOZ [21]. However, having patients with electrodes implanted for long periods of time, and requiring the monitoring of multiple seizure events over many weeks carries the risk of infection, sudden death, trauma and cognitive deficits from having repeated seizures. This contributes to the large cost of epilepsy monitoring [2]. If a candidate iEEG marker could be found that is able to provide strong localizing evidence using only interictal data, then it would significantly reduce invasive monitoring time [47].
Neural fragility is an EEG marker that can also further advance our knowledge of neural mechanisms of seizure generation that will then drive more effective interventions. For example, fragility can be used to identify pathological tissue that are then removed and tested in vitro for abnormal histopathological structure [28, 27]. Knowledge of structural abnormalities may inform new targeted drug treatments. In the future, specific fragility patterns can be correlated with specific pathological substrates. Likely, the specific pathological substrates will have different therapeutic approaches. As an example, epilepsy caused by focal cortical dysplasia is treated with focal surgical resection, but post-encephalitic epilepsy may have a better therapeutic response with immunosuppressants and steroids.
Finally, neural fragility may have broader implications in understanding how underlying brain network dynamics change during intervention (e.g. drugs or electrical stimulation). Fragility analysis can be applied as a method of assessing the efficacy of specific drug trials to specific pathological groups, which include not only epilepsy but other neurological conditions such as Alzheimer’s’s disease or the spectrum of dementias. Commonly, the current optimal criteria to recognize therapeutic success in many neurological conditions is purely clinical, but clinical responses are not immediate. This delay in recognizing the appropriate drug and adequate therapeutic doses is highly detrimental. Computational methods as fragility measuring EEG before and after drug administration could provide an additional criteria for drug responses. This quantitative measurement can be immediate, guiding the treating physician to the correct treatment without delays and unnecessary drug trials. Furthermore, if neural fragility could be accurately obtained from non-invasive tests or from permanently implanted devices, the current fragility of the network could be used as a surrogate marker of patient’s current clinical state. As such, the changes in the fragility could be used as a proxy for improvement or recurrences that occur as medication doses (or other treatments, such as Keto Diet [50]) are changed over time.
10. Methods
All data were acquired with approval of local Institutional Review Board (IRB) at each clinical institution: UMMC by IRB of the University of Maryland School of Medicine, UMH by University of Miami Human Subject Research Office - Medical Sciences IRB, NIH by the National Institute of Health IRB, JHH by Johns Hopkins IRB and CClinic by Cleveland Clinic Institutional Review Board. Informed consent was given at each clinical center. The acquisition of data for research purposes was completed with no impact on the clinical objectives of the patient stay. Digitized data were stored in an IRB-approved database compliant with Health Insurance Portability and Accountability Act (HIPAA) regulations.
10.1. Dataset collection
iEEG data from 91 DRE patients who underwent intracranial EEG monitoring, which included either electrocorticography (ECoG), or depth electrodes with stereo-EEG (SEEG) were selected from University of Maryland Medical Center (UMMC), University of Miami Jackson Memorial Hospital (UMH), National Institute of Health (NIH), Johns Hopkins Hospital (JHH), and the Cleveland Clinic (CClinic). Patients exhibiting the following criteria were excluded: no seizures recorded, pregnant, sampling rate less than 250 Hz, previous surgeries more then 6 months before, and no surgery performed (possibly from SOZ localizations in eloquent areas). We define successful outcomes as seizure free (Engel class I and ILAE scores of 1 and 2) at 12+ months post-op and failure outcomes with seizure recurrence (Engel classes 2-4) [51, 52, 29]. Of these 91 patients, 44 experienced successful outcomes and 47 had failed outcomes (age at surgery = 31.52 +/− 12.32 years) with a total of 462 seizures (seizure length = 97.82 +/− 91.32 seconds) and 14703 total number of recording electrodes (159.82 +/− 45.42 per subject). Decisions regarding the need for invasive monitoring and the placement of electrode arrays were made independently of this work and part of routine clinical care. Data collection and labeling of patients were done retrospectively and thus blind to the authors conducting data analysis (i.e. AL, CH, SVS), but not with respect to the corresponding clinicans. Data analysis were not performed blind to the conditions of the experiments. Sample sizes were determined by attempting to gather as many patients as possible from multiple clinical centers (see Table 1 for summary of all patients gathered). When possible, we sought to have a balance of lesional, temporal, extra-temporal and multi-focal epilepsy patients, along with a relatively proportional number of success and failed epilepsy surgery patients (success = seizure freedom, failure = seizure recurrence). We collected 100 patients and then removed 9 due exclusion criterion. No statistical methods were used to pre-determine sample sizes but our sample sizes are larger then those reported in previous publications [11].
Our clinician team categorized patients by surgical outcome, Engel class and ILAE score. In addition, we categorized patients by their clinical complexity (CC) as follows: (1) lesional, (2) focal temporal, (3) focal extratemporal, and (4) multi-focal (Figure 3) [29, 21]. Each of these were categorized based on previous outcome studies that support this increasing level of localization difficulty. Lesional patients have success rates of ~70%, experiencing the highest rate of surgical success because the lesions identified through MRI are likely to be part of the SOZ [53]. Localization and surgical success are more challenging patients with non-lesional MRI, with average surgical success rates in temporal, extratemporal and multi-focal epilepsy of ~60%, ~45% and ~30%, respectively [31, 32]. Patients that fit into multiple categories were placed into the more complex category. Next, clinicians identified electrodes that they hypothesized as SOZ. In general, this was a subset of the resected region for all patients, unless otherwise noted. The epileptologists define the clinically annotated SOZ as the electrodes that participated the earliest in seizures. Every patient’s SOZ was labeled by 1-3 trained epileptologists (depending on the center). The corresponding SOZ complement that we define, or SOZC, are the electrodes that are not part of the SOZ. For more detailed information regarding each patient, see Supplemental clinical data summary Excel file and Table 1.
At all centers, data were recorded using either a Nihon Kohden (Tokyo, Japan) or Natus (Pleasanton, USA) acquisition system with a typical sampling rate of 1000 Hz (for details regarding sampling rate per patient, see Supplementary table). Signals were referenced to a common electrode placed subcutaneously on the scalp, on the mastoid process, or on the subdural grid. At all centers, as part of routine clinical care, up to three board-certified epileptologists marked the electrographic onset and the termination of each seizure by consensus. The time of seizure onset was indicated by a variety of stereotypical electrographic features which included, but were not limited to, the onset of fast rhythmic activity, an isolated spike or spike-and-wave complex followed by rhythmic activity, or an electrodecremental response. Each snapshot available for a patient was clipped at least 30 seconds before and after the ictal event. We discarded electrodes from further analysis if they were deemed excessively noisy by clinicians, recording from white matter, or were not EEG related (e.g. reference, EKG, or not attached to the brain) which resulted in 97.23 +/− 34.87 (mean +/− std) electrodes used per patient in our analysis. Data was initially stored in the form of the European Data Format (EDF) files [54]. We preprocessed data into the BIDS-iEEG format and performed processing using Python3.6, Numpy, Scipy, MNE-Python and MNE-BIDS [55, 56, 57, 58, 59, 60, 61]. Figures were generated using Matplotlib and Seaborn [62, 63, 64]. Statistical analyses were performed with mlxtend, pingouin, dabest, scipy and scikit-learn [65, 66, 67, 61, 68].
10.2. Preprocessing of data
Every dataset was notch filtered at 60 Hz (with a cutoff window of 2 Hz) and bandpass filtered between 0.5 and the Nyquist frequency with a fourth order Butterworth filter. A common average reference was applied to remove any correlated noise. EEG sequences were broken down into sequential windows and the features were computed within each window (see Methods - Neural fragility of iEEG network, Baseline features - spectral features and Baseline features - graph analysis of networks for details). Each proposed feature representation produces a value for each electrode for each separate window, and results in a full spatiotemporal heatmap when computed over sequential windows with electrodes on the y-axis, time on the x-axis and feature value on the color axis. In total, we computed 20 different baseline feature representations from the iEEG data: 6 frequency power bands, 7 eigenvector centralities (one for each frequency band coherence connectivity matrix and one for a correlation connectivity matrix), 7 degrees (one for each frequency band coherence connectivity matrix and one for a correlation connectivity matrix). Values at each window of time were normalized across electrodes to values that could range from 0 up to at most 1, to allow for comparison of relative feature value differences across electrodes over time; the higher a normalized feature, the more we hypothesized that electrode was part of the SOZ [16].
10.3. Neural fragility of iEEG network
When one observes iEEG data during interictal, or preictal periods, activity recorded from each channel is noisy and hovers around a baseline value. In contrast, when one observes iEEG data during a seizure event, activity (i) grows in amplitude, (ii) oscillates, and (iii) spreads in the brain. From a dynamical systems perspective, the iEEG network has switched from a stable (non-seizure) to an unstable (seizure) network. The only difference between the iEEG networks in Figure 2 is the connection strengths representing the dynamical interactions between a few channels, i.e., the SOZ. Our conjecture is that small changes in connection strengths at SOZ nodes cause an imbalance in inhibitory and excitatory connectivity between brain regions. Either inhibition is decreased and/or excitation is increased; thus, if the SOZ is perturbed then over excitation can occur manifesting in a seizure.
To compute fragility heatmaps from iEEG recordings, we first constructed simple linear models as described above and in equation 1.
(1) |
Since each observation, , has dimension d (number of channels), we would like to formulate a least-squares estimation procedure with n > d samples. We choose n to represent a 250 ms iEEG window. We then have the following representation of :
(2) |
and the following representation for
(3) |
The least-squares will now seek to fit a linear operator A such that:
(4) |
This linear operator representation of the dynamical system has connections to Koopman operator theory [69] and Dynamic Mode Decomposition in Fluid Mechanics [70]. We seek to approximate the inherently nonlinear iEEG dynamics within a small window of time using a finite-dimensional approximation of the Koopman operator using the observables (i.e. x(t)) themselves. We specifically used least-squares algorithm with a 10e-5 l2-norm regularization to ensure that the model identified was stable (with absolute value of eigenvalues ≤ 1) as in [17, 16]. Then, we slid the window 125 ms and repeated the process over the entire data sample, generating a sequence of linear network models in time as in Extended Data Figure 1).
We systematically computed the minimum perturbation required for each electrode’s connections (Extended Data Figure 1) to produce instability for the entire network as described in [16]. This is represented in equation 5, where the Δi is the desired column perturbation matrix for channel i, and the is the desired radii to perturb one single eigenvalue to.
(5) |
More specifically, we compute a structured perturbation matrix, such that:
(6) |
where each is the actual column perturbation vector. The intuition for using this type of structured perturbation is described in the main paper in section Description of Neural Fragility. We demonstrate how to solve for this using least-squares in [16]. The electrodes that were the most fragile were hypothesized to be related to the SOZ in these epilepsy networks (seen as the dark-red color in the turbo colormap in Figure 4).
10.4. Baseline features - spectral features
We constructed spectral-based features from frequency bands of interest by applying a multi-taper Fourier transform over sliding windows of data with a window/step size of 2.5/0.5 seconds [12]. We required relatively longer time windows to accurately estimate some of the lower frequency bands. Each EEG time series was first transformed into a 3-dimensional array (electrodes X frequency X time), and then averaged within each frequency band to form six different spectral feature representations of the data. We break down frequency bands as follows:
Delta Frequency Band [0.5 - 4 Hz]
Theta Frequency Band [4 - 8 Hz]
Alpha Frequency Band [8 - 13 Hz]
Beta Frequency Band [13 - 30 Hz]
Gamma Frequency Band [30 - 90 Hz]
High-Gamma Frequency Band [90 - 300 Hz]
The High-Gamma frequency band includes frequencies of “ripples” - a category of HFOs.
10.5. Baseline features - graph analysis of networks
We computed a time domain model using Pearson correlation (equation 7) and a frequency domain model using coherence (equation 8). We computed the connectivity matrix using MNE-Python and used the default values [58]. In the equations, i, j) are the electrode indices, Cov is the covariance, σ is the standard deviation, f is the frequency band, and G is cross-spectral density. These connectivity models attempt to capture linear correlations either in time, or in a specific frequency band, but are not dynamical system representations of the data (i.e. x(t + 1) = Ax(t)). For each network-based feature, a sliding window/step size of 2.5/0.5 seconds were used, resulting in a sequence of network matrices over time resulting in 3-dimensional arrays (electrodes X electrodes X time) [11, 12, 13].
(7) |
(8) |
From each network matrix, we computed the eigenvector centrality [12, 11], and the degree features of the network for each electrode across time. Since coherence and pearson correlation are symmetric matrices, the in-degree is equivalent to the out-degree. Centrality describes how influential a node is within a graph network. The degree is the weighted sum of the connections that connect to a specific node. Both features are potential measures that attempt to capture the influence of a specific electrode within an iEEG network, through the lens of graph theory.
10.6. Experimental design
We tested if the neural fragility representation of iEEG data localized the SOZ better compared to other proposed features, compared to clinicians and compared to chance. Electrodes with extreme feature values deviating from the average were hypothesized as part of the SOZ.
To compute a probability of successful surgical outcome for each patient, we trained a non-parametric machine learning classifier, a Random Forest classifier to output a probability value that the surgery was a success. We input the distribution of feature values in the clinically SOZ and the SOZC. We considered a set of hyperparameters that we then performed evaluation over a 10-fold nested cross-validation scheme. We then performed statistical analysis on the final classification performance to determine the most robust feature representation.
Pooled Patient Analysis
First, we analyzed the difference in the distributions of neural fragility between SOZ and the SOZC. We pooled all patients together, stratified by surgical outcome and compared the neural fragility distributions using a one-sided Mann-Whitney U test (Success pvalue = 3.326e-70, Fail pvalue = 0.355; Extended Data Figure 3). This suggested that there is some effect on average where fragility is higher in the SOZ for successful outcomes, so we next looked at the distributions per patient’s seizure snapshot around seizure onset. In Extended Data Figure 4, success outcome patients have a higher neural fragility in the SOZ. This effect is seen when pooling patients across all centers as well, where neural fragility is either i) higher before the seizure onset, or ii) has a marked difference starting at seizure onset (Extended Data Figure 5).Next, we performed a classification experiment (Figure 3) that would determine the robustness of the neural fragility representation at the patient level benchmarked against 20 other features.
Non-parametric Decision-Tree Classifier
To determine the value of a feature representation of the iEEG data, we posed a binary classification problem where the goal would be to determine the surgical outcome (success or failure) for a particular patient’s spatiotemporal heatmap. Each spatiotemporal heatmap was split into its SOZ and SOZC set (FSOZ and FSOCC in Extended Data Figure 1).Then each set of electrodes’ feature distributions were summarized with its quantile statistics over time, resulting in twenty signals: ten quantiles from 10-100 of the SOZ and ten quantiles of the SOZC over time. We used a Random Forest (RF) classifier [71]. Specifically, it was a variant known as the Structured Random Forest, or Manifold Random Forests [72, 73]. The manifold RF approach allows one to encode structural assumptions on the dataset, such as correlations in time of the fed in data matrix. The input data matrix for each RF is a multivariate time series that summarizes the quantile statistics of the SOZ and SOZC over time. As a result, we obtained better results using the manifold RF because it took advantage of local correlations in time, rather than treating all values in the data matrix input as independent as done in a traditional RF model. This approach allowed our classifiers to learn faster with less data, compared to treating all inputs as independent in the traditional RF model. For more information on how manifold RF improves on traditional RF, we refer readers to [72, 73]. For every model, we used the default parameters from scikit-learn and the rerf package [68, 73]: 500 estimators, max depth is None, minimum samples split is 1, max features is auto, feature combinations is 1.5, image height of 20 (20 quantiles total), patch height max of 4, patch height minimum of 1, patch width max of 8, patch width minimum of 1. The output of the decision tree classifier is:
where θ are the trained RF parameters and FSOZ are the heatmap values for the SOZ and FSOZC are the heatmap values for the SOZC for a specific feature representation F (shown in Extended Data Figure 1). is the function we are trying to estimate, which predicts the probability of successful surgical outcome for a given feature heatmap.
Hyperparameters
When looking at iEEG data, clinicians inherently select windows of interest due to external and prior information about when the seizure clinically manifests. Similarly, we consider a window of 10 seconds before seizure onset to the first 5% of the seizure event. This window was chosen apriori to analysis, and we repeated the analyses with slightly varying windows, but the results were consistent. To provide further contrast to the spatiotemporal heatmaps, we considered also a threshold between 0.3 and 0.7 (spaced by 0.1) that would be applied to the heatmap such that values below were set to 0. This is analagous to clinicians being able to look at a iEEG data and hone in only on areas that are extreme relative to the rest. We selected a fixed threshold through nested cross-validation, where thresholds are selected on the train/validation set, and then performance is measured on the held out test set.
Structured Heatmap Input
To train a RF classifier on patients with varying number of channels, we structured the spatiotemporal heatmap inputs to summarize the distributions of the feature values in the SOZ and SOZC sets. We converted these into a vector of 20 quantile values (10 quantiles for SOZ and 10 for SOZC), taken evenly from the 10th to the 100th quantiles. This forms a data matrix per heatmap, which summarizes the SOZ and SOZC distributions over time, fixed around the seizure onset time.
Nested Cross-Validation Feature Evaluation
It is common practice when building a supervised machine learning model to incorporate cross-validation (CV) where the dataset is split into a training and testing set. However, when one has hyperparameters (in addition to the machine learning model parameters) it is crucial to control against over-fitting to the test dataset [74]. Due to our hyperparameter selection of the optimal heatmap thresholds, we used a nested CV scheme, where we split the dataset into a training, validation and a “lock-box” dataset, where the hyperparameters were tuned on an inner CV (70% of the dataset) with the training and validation data, and then performance was evaluated on the lock-box dataset. All features were optimized separately, so that their hyperparameters were optimal for that feature. We computed the discrimination statistic known as the area under the curve (AUC). We repeated the nested-cross validation 10 times, resulting a 10-fold CV. Statistical analysis was done using nested cross-validation, splitting training, validation and testing groups based on subjects, while sampling proportionally the different epilepsy etiologies. We set aside 60% of subjects for training, 10% for validation and then 30% as a held-out test set. We used a heuristic of attempting to sample approximately randomly from clinical complexity 1-4 as defined in the paper. This heuristic attempted to place at least 25% of the subjects in the training, validation and held-out test set for each clinical complexity. Besides this covariate, the training, validation and held-out test set were determined randomly. We found that this performed slightly better for all features compared to random sampling. We performed patient-level CV, ensuring that no patient was in multiple splits of the dataset (i.e. any one patient is only in the train, test, or lock-box dataset).
Statistical analysis - Hypothesis Testing
Success probability values were computed from the RF classifiers trained on the spatiotemporal feature representations of the iEEG data (fragility, spectral features, and graph metrics from correlation and coherence graphs) resulting in a distribution of probabilities for each feature.
To compare the RF model performance across the 20 proposed features of fragility, spectral power and graph metrics, we computed Cohen’s D effect size differences between the groups of interest, and then Mann-Whitney U tests for unpaired data, and paired Wilcoxon rank-sum tests for paired data [65]. We corrected for any multiple comparisons using the Holm-Bonferroni step-down method. In some cases, we also present the Likelihood Q-ratio test (LQRT) results to show that the results are robust to statistical test chosen. The LQRT has been shown to be more robust compared to both the Mann-Whitney U-test and t-test in the presence of noise [75, 76]. LQRT utilizes a bootstrapping procedure, so it’s resolution is 0.001 (i.e. if it produces a p-value of 0, then that means it is < 0.001). Since the Mann-Whitney, Wilcoxon and LQRT tests do not rely on a parametric assumption of normality, no explicit parametric assumptions were made. In certain cases, where we report results of a t-test as well, we assumed that the data distribution might be normal, but this was not formally tested.
We compared the success probabilities stratified by different clinical covariates: surgical outcome, Engel class, clinical complexity, handedness, gender, onset age, and surgery age. We then estimated the effect size differences between distributions in the form of a Cohen’s d statistic. Cohen’s d was estimated using a non-parametric bootstrapping on the observed data with 5000 re-samples used to construct a 95% confidence interval [65]. The null hypothesis of our experimental setup was that the success probabilities came from the same population. The alternative hypothesis was that the populations were different (i.e. a feature could distinguish success from failed outcomes, or between different Engel classes).
10.7. Feature evaluation using predicted probability of successful surgery
The fragility and all baseline features proposed generated a spatiotemporal heatmap using EEG snapshots of ictal data. To compare spatiotemporal heatmaps across features, we computed a probability of success (i.e. a confidence score) that is hypothesized to be high for success outcomes and low for failures for “good” features. We expected that fragility would follow a trend of decreasing confidence as CC, Engel score and ILAE score increase. For each clinical covariate group, we measured the effect size difference via bootstrapped sampling, and the statistical p-value between the distributions. We hypothesized that: i) fragility would have an effect size difference significantly different from zero when comparing success vs failed outcomes, ii) in addition, this effect size would correlate with meaningful clinical covariates, such as CC and Engel class and iii) both the effect size and p-value would be better than the proposed baseline features.
The higher the probability of success (closer to 1), the more likely the feature indicated a successful surgery, and the lower it was (closer to 0), the more likely the feature indicated a failed surgery. To compute this value, we first partition the heatmap into a SOZ and SOZC. This forms the two sets of signals that represent the spatiotemporal feature values of the SOZ set vs the SOZC set of electrodes. Then we take windows of interest, where clinicians find most valuable for SOZ localization: the period before seizure and right after seizure onset, and performing a nested CV of RF models. The efficacy of each proposed feature is evaluated based on how well the trained RF model is able to predict the surgical outcomes. We tested our hypotheses stated above by computing a probability of success from each feature heatmap for each patient, and estimated the distribution differences of the CS between various clinical covariates.
10.8. Spatiotemporal feature heatmap interpretability
To determine how valid the output probability values are, we first computed a calibration curve, which told us how well-calibrated the probability values were. We furthermore compared the calibration curves across neural fragility, spectral features and graph metrics. With relatively good discrimination measured by the AUC and good calibration, we were able to then compare the CS across clinical complexity (CC) scores, Engel scores and ILAE scores. Since epilepsy and reasons for failed surgeries are so complex, these are clinical methods for stratifying patient groups based on observed etiology. We analyzed how the success probability differs across each of these categories. In addition, we looked at how the success probabilities might differ across other clinical variables, such as sex (M vs F), handedness (R vs L), epilepsy onset age (years), and age during surgery (years).
Qualitatively reading off a spatiotemporal heatmap is highly interpretable as one can match raw iEEG segments to certain time periods in the heatmap. For every patient, the time-varying quantile signals (from 10th to the 100th) of the SOZ and SOZC were computed for every iEEG heatmap, as specified in Methods - Experimental design. We used permutations feature importance sampling to obtain a relative weighting of each signal over time and how important it was in allowing the RF to correctly determine surgical outcome. This was visualized as a heatmap showing the mean and std importance scores of the SOZ and SOZC statistics, as shown in Extended Data Figure 7).
In addition, we claim that human interpretability of the spatiotemporal heatmap relies on contrast between the SOZ and SOZC regions in successful outcomes. Using the results of the feature importance permutation test, we computed for every heatmap, an interpretability ratio, which was defined as:
where I is the interpretability ratio for a specific subject’s feature heatmap F. FSOZ(90th) is the 90th quantile and up feature values of the SOZ, and FSOCC (90th) is for the SOZC. This ratio is then stratified according to surgical outcome, which we expect higher ratios for successful outcomes and lower ratios for failed outcomes. We quantified this difference using Cohen’s D effect size and Mann Whitney U-test Pvalues (see Methods - Statistical analysis - Hypothesis Testing).
10.9. Fragility heatmaps are more interpretable than all other EEG feature maps
We can parse the RF model to determine important aspects of the spatiotemporal heatmap that are important in predicting surgical outcome. To do so, we perform permutations on the heatmap inputs to the RF models to measure their relative importances over space and time. We observed that the AUC metric was affected primarily by a combination of the highest (90th quantiles and up) neural fragility values of both the SOZ and SOZC contacts (Extended Data Figure 7). The SOZC neural fragility as much as 10 seconds before seizure onset impacted the AUC as did neural fragility of the SOZ right at seizure onset. For both the SOZ and the SOZC fragility distributions, the 80th quantile and below did not contribute to the predictive power. In different patients across different clinical centers, the practice of marking seizure onset times can vary due to varying methodologies [21]. As a result, in the SOZC, there was some variation in terms of which time point in the heatmap mattered most.
Although strong predictive capabilities of an EEG marker of the SOZ are promising, it is also important that the marker be presented in an interpretable manner to clinicians. We next show how fragility heatmaps are the most interpretable over all baseline features. In Extended Data Figure 9a, there are two heatmaps computed for the same seizure event in Patient_01: one is a beta band map (left) and one is a neural fragility map (right). Both maps are normalized across channels and both are computed with similar sliding windows. However, less contacts “stand-out” as pathological in the beta-band heatmap before seizure onset, with the majority of the map being different shades of blue. In contrast, in the fragility heatmap, one contact (ATT1 from Figure 4) is fragile the entire duration before seizure onset (solid white line), and then a few more contacts become fragile after the electrographic onset of the seizure. These fragile areas that “pop-out” in the heat-map as red-hot areas occur in the clinically annotated SOZ and this patient had a successful surgery.
To quantify interpretability, we compute an interpretability ratio: the ratio of the feature values in the 90th quantile between the SOZ and the SOZC over the section of data used by the feature’s RF model. This measures the contrast that one sees between the extreme values in the SOZ versus the extreme values they see in the SOZC. The larger the ratio, then the more contrast the map will have. In Extended Data Figure 9b, we show that the effect size difference between successful and failed outcome of this interpretability ratio is largest in neural fragility when compared to all baseline features. It is well-known that RF models are scale-invariant [71], so it is plausible that there are portions of heatmaps that distinguish one channel from another that can be parsed out via a decision tree and not the naked eye, which leads to high AUC in the beta band representation in Figure 5. For example a decision tree can discriminate between 0.3000 and 0.2995, which on a normalized color-scale is difficult to parse by visual inspection. Neural fragility on the other hand, shows marked differences between the clinically annotated SOZ (red electrodes on the y-axis) and the actual fragility values even before seizure onset.
10.10. Data availability
We also released the raw iEEG data for patients from NIH, UMH, UMMC, and JHH in the OpenNeuro repository in the form of BIDS-iEEG (https://openneuro.org/datasets/ds003029). Due to restrictions on data sharing from CClinic, we were unable to release the raw iEEG data that we received from this site. Dataset from CClinic is available upon request from authors at the CClinic.
10.11. Code availability
All main figures of the manuscript can be reproduced using Gigantum at https://gigantum.com/adam2392/neural-fragility-ictal-study-figures [77]. We include a jupyter notebook written in Python to help reproduce figures. An example of neural fragility being run on Patient_01 can be shared upon request due to licensing restrictions.
The neural fragility algorithm has been implemented in a FDA 510k approved software medical device (current 510k number: K201910). More information will be available at www.neurologicsolutions.co. Otherwise, please contact corresponding authors for more information regarding a clinical demonstration.
Extended Data
Supplementary Material
5. Acknowledgements
AL is supported by NIH T32 EB003383, the NSF GRFP (DGE-1746891), the Arcs Chapter Scholarship, Whitaker Fellowship and the Chateaubriand Fellowship. SVS is supported by NIH R21 NS103113, the Coulter Foundation, Maryland Innovation Initiative, US NSF Career Award 1055560 and the Burroughs Wellcome Fund CASI Award 1007274. EJ is supported by NIA K23 AG063899. IC was partially supported by NIH R25NS108937-02. KZ and SI are supported by the Intramural Research Program of the National Institute of Neurological Disorders and Stroke. SC is supported by 1UL1TR003098-01 and the UMB Institute for Clinical and Translational Research. JHopp is supported by NINDS 1 UO1 NS08803401 - Established Status Epilepticus Treatment Trial (ESETT), MII TA Grant 20180828 TEDCO - EpiWatch: A Mobile Health Application for Epilepsy Management and U10 NS058932-01 NIH NINDS - “Neurological Emergencies Treatment Trials (NETT) Network Clinical/Site Hub”. Computational resources were also provided by the Maryland Advanced Research Computing Center (MARCC). The authors would like to thank Carey Priebe for useful discussions on statistical analyses, Joshua Vogelstein for introducing AL to the concepts of Structured Random Forests, Sarah Kim and Rachel June Smith for very helpful reviews of the manuscript and Macauley Breault for help in creating brain figures. Finally, the authors would like to thank the editor and anonymous reviewers for their insightful comments and reviews.
Footnotes
Competing Interests Statement
The authors declare the following competing interests: AL, SVS and JM have equity in a startup, Neurologic Solutions Co., related to epilepsy data analysis. WA is on the Board of Advisors of Longeviti Neuro Solutions and a Compensated Consultant at Globus Medical. JHopp reports the following disclosures: Personal compensation: Royalties from publishing – UpToDate, Inc., Honoraria from speaking engagements - J. Kiffin Penry Epilepsy Education Programs, Expert witness consultation: Gerolamo McNulty Divis & Lewbart A Professional Corporation Attorneys At Law Goodell, Devries, Leech & Dann, LLP. All other authors declare no competing interests.
4 References
- [1].Berg AT Identification of Pharmacoresistant Epilepsy. Neurologic Clinics 27, 1003–1013 (2009). [DOI] [PMC free article] [PubMed] [Google Scholar]
- [2].WHO & Organization, W. H. Epilepsy (2019).
- [3].Jobst BC Consensus over individualism: Validation of the ILAE definition for drug resistant epilepsy Epilepsy Currents (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
- [4].Begley CE & Durgin TL The direct cost of epilepsy in the United States: A systematic review of estimates Epilepsia (2015). [DOI] [PubMed] [Google Scholar]
- [5].Jehi L The epileptogenic zone: Concept and definition Epilepsy Currents (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- [6].Penfield W Epileptogenic lesions. Acta neurologica et psychiatrica Belgica 56, 75–88 (1956). [PubMed] [Google Scholar]
- [7].Jobst BC, Darcey TM, Thadani VM & Roberts DW Brain stimulation for the treatment of epilepsy: Brain Stimulation in Epilepsy. Epilepsia 51, 88–92 (2010). [DOI] [PubMed] [Google Scholar]
- [8].Scharfman HE The neurobiology of epilepsy Current Neurology and Neuroscience Reports (2007). [DOI] [PMC free article] [PubMed] [Google Scholar]
- [9].Gliske SV et al. Universal automated high frequency oscillation detector for real-time, long term EEG. Clinical Neurophysiology 127, 1057–1066 (2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
- [10].Shah P et al. Local structural connectivity directs seizure spread in focal epilepsy. Preprint at https://www.biorxiv.org/content/10.1101/406793v1.full (2018). [Google Scholar]
- [11].Li A et al. Using network analysis to localize the epileptogenic zone from invasive EEG recordings in intractable focal epilepsy. Network Neuroscience 02, 218–240 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- [12].Burns SP et al. Network dynamics of the brain and influence of the epileptic seizure onset zone. Proceedings of the National Academy of Sciences 111, E5321–E5330 (2014). [DOI] [PMC free article] [PubMed] [Google Scholar]
- [13].Yaffe RB et al. Physiology of functional and effective networks in epilepsy. Clinical Neurophysiology 126, 227–236 (2015). [DOI] [PubMed] [Google Scholar]
- [14].Khambhati AN, Davis KA, Lucas TH, Litt B & Bassett DS Virtual Cortical Resection Reveals Push-Pull Network Control Preceding Seizure Evolution. Neuron 91, 1170–1182 (2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
- [15].Bassett DS & Sporns O Network neuroscience. Nature neuroscience 20, 353–364 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- [16].Li A, Inati S, Zaghloul K & Sarma S Fragility in Epileptic Networks : the Epileptogenic Zone. In American Control Conference, 1–8 (2017). [Google Scholar]
- [17].Li A et al. Linear time-varying model characterizes invasive EEG signals generated from complex epileptic networks. In Proceedings of the Annual International Conference of the IEEE Engineering in Medicine and Biology Society, EMBS, 1, 2802–2805 (IEEE, 2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- [18].Sritharan D & Sarma SV Fragility in Dynamic Networks: Application to Neural Networks in the Epileptic Cortex. Neural Computation 26, 2294–2327 (2014). [DOI] [PubMed] [Google Scholar]
- [19].Jette N, Reid AY & Wiebe S What are the limitations of medical therapy for epilepsy? Surgical management of epilepsy Key points. CMAJ 186, 186 (2014). [DOI] [PMC free article] [PubMed] [Google Scholar]
- [20].Engel J et al. Early surgical therapy for drug-resistant temporal lobe epilepsy: A randomized trial. JAMA - Journal of the American Medical Association 307, 922–930 (2012). [DOI] [PMC free article] [PubMed] [Google Scholar]
- [21].Sheikh S, Nair D, Gross R & Gonzalez Martinez J Tracking a changing paradigm and the modern face of epilepsy surgery: A comprehensive and critical review on the hunt for the optimal extent of resection in mesial temporal lobe epilepsy. Epilepsia 60, 1768–1793 (2019). [DOI] [PubMed] [Google Scholar]
- [22].Crino PB et al. Increased expression of the neuronal glutamate transporter (EAAT3/EAAC1) in hippocampal and neocortical epilepsy. Epilepsia 43, 211–218 (2002). [DOI] [PMC free article] [PubMed] [Google Scholar]
- [23].González-Martínez JA, Ying Z, Prayson R, Bingaman W & Najm I Glutamate clearance mechanisms in resected cortical dysplasia: Laboratory investigation. Journal of Neurosurgery 114, 1195–1202 (2011). [DOI] [PubMed] [Google Scholar]
- [24].Noebels JL The biology of epilepsy genes Annual Review of Neuroscience (2003). [DOI] [PubMed] [Google Scholar]
- [25].Bradford HF Glutamate, GABA and epilepsy Progress in Neurobiology (1995). [DOI] [PubMed] [Google Scholar]
- [26].Traub RD et al. A possible role for gap junctions in generation of very fast EEG oscillations preceding the onset of, and perhaps initiating, seizures. Epilepsia 42, 153–170 (2001). [DOI] [PubMed] [Google Scholar]
- [27].DeFelipe J Chandelier cells and epilepsy. Brain 122, 1807–1822 (1999). [DOI] [PubMed] [Google Scholar]
- [28].Jin X, Prince DA & Huguenard JR Enhanced Excitatory Synaptic Connectivity in Layer V Pyramidal Neurons of Chronically Injured Epileptogenic Neocortex in Rats. Journal of Neuroscience 26, 4891–4900 (2006). [DOI] [PMC free article] [PubMed] [Google Scholar]
- [29].Bulacio JC, Chauvel P & McGonigal A Stereoelectroencephalography: Interpretation Journal of Clinical Neurophysiology (2016). [DOI] [PubMed] [Google Scholar]
- [30].Couronné R, Probst P & Boulesteix AL Random forest versus logistic regression: A large-scale benchmark experiment. BMC Bioinformatics 19, 270 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- [31].Sheikh S et al. Redefining success in epilepsy surgery: The importance of relative seizure reduction in patient reported quality of life. Epilepsia 2078–2085 (2019). [DOI] [PubMed] [Google Scholar]
- [32].Vakharia VN et al. Getting the best outcomes from epilepsy surgery. Annals of Neurology (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- [33].Varatharajah Y et al. Integrating artificial intelligence with real-time intracranial EEG monitoring to automate interictal identification of seizure onset zones in focal epilepsy. Journal of Neural Engineering 15 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- [34].Laxer KD et al. The consequences of refractory epilepsy and its treatment Epilepsy & Behavior (2014). [DOI] [PubMed] [Google Scholar]
- [35].McIntosh AM et al. Long-term seizure outcome and risk factors for recurrence after extratemporal epilepsy surgery. Epilepsia 53, 970–978 (2012). [DOI] [PubMed] [Google Scholar]
- [36].Yardi R et al. Long-term outcomes of reoperations in epilepsy surgery. Epilepsia 61, 465–478 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- [37].Jeppesen J, Beniczky S, Johansen P, Sidenius P & Fuglsang-Frederiksen A Detection of epileptic seizures with a modified heart rate variability algorithm based on Lorenz plot. Seizure 24, 1–7 (2015). [DOI] [PubMed] [Google Scholar]
- [38].Chung J et al. Statistical Connectomics. Preprint at https://osf.io/ek4n3/https://osf.io/ek4n3 (2020). [Google Scholar]
- [39].Fedele T et al. Resection of high frequency oscillations predicts seizure outcome in the individual patient. Scientific Reports 7, 13836 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- [40].Haegelen C et al. High-frequency oscillations, extent of surgical resection, and surgical outcome in drug-resistant focal epilepsy. Epilepsia 54, 848–857 (2013). [DOI] [PMC free article] [PubMed] [Google Scholar]
- [41].Cuello-Oderiz C et al. Value of ictal and interictal epileptiform discharges and high frequency oscillations for delineating the epileptogenic zone in patients with focal cortical dysplasia. Clinical Neurophysiology 129, 1311–1319 (2018). [DOI] [PubMed] [Google Scholar]
- [42].Höller Y et al. High-frequency oscillations in epilepsy and surgical outcome. A meta-analysis. Frontiers in Human Neuroscience 9, 574 (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
- [43].Spring AM et al. Interrater reliability of visually evaluated high frequency oscillations. Clinical Neurophysiology 128, 433–441 (2017). [DOI] [PubMed] [Google Scholar]
- [44].Höller P, Trinka E & Höller Y High-Frequency Oscillations in the Scalp Electroencephalogram: Mission Impossible without Computational Intelligence Computational Intelligence and Neuroscience (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- [45].Gloss D, Nevitt SJ & Staba R The role of high-frequency oscillations in epilepsy surgery planning Cochrane Database Systematic Reviews (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- [46].Chang WC et al. Loss of neuronal network resilience precedes seizures and determines the ictogenic nature of interictal synaptic perturbations. Nature Neuroscience 21, 1742–1752 (2018). [DOI] [PubMed] [Google Scholar]
- [47].Kramer MA & Cash SS Epilepsy as a disorder of cortical network organization Neuroscientist (2012). [DOI] [PMC free article] [PubMed] [Google Scholar]
- [48].Meador KJ, Kapur R, Loring DW, Kanner AM & Morrell MJ Quality of life and mood in patients with medically intractable epilepsy treated with targeted responsive neurostimulation. Epilepsy and Behavior 45, 242–247 (2015). [DOI] [PubMed] [Google Scholar]
- [49].Loring DW, Kapur R, Meador KJ & Morrell MJ Differential neuropsychological outcomes following targeted responsive neurostimulation for partial-onset epilepsy. Epilepsia 56, 1836–1844 (2015). [DOI] [PubMed] [Google Scholar]
- [50].D’Andrea Meira I et al. Ketogenic diet and epilepsy: What we know so far. Frontiers in Neuroscience 13 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- [51].Hawasli AH, Bandt SK, Hogan RE, Werner N & Leuthardt EC Laser ablation as treatment strategy for medically refractory dominant insular epilepsy: Therapeutic and functional considerations. Stereotactic and Functional Neurosurgery 92, 397–404 (2014). [DOI] [PMC free article] [PubMed] [Google Scholar]
- [52].Gonzalez-Martinez J & Lachhwani D Stereoelectroencephalography in children with cortical dysplasia: technique and results. Child’s Nervous System 30, 1853–1857 (2014). [DOI] [PubMed] [Google Scholar]
- [53].See SJ et al. Surgical outcomes in patients with extratemporal epilepsy and subtle or normal magnetic resonance imaging findings. Neurosurgery 73, 68–76 (2013). [DOI] [PubMed] [Google Scholar]
- [54].Kemp B, Värri A, Rosa AC, Nielsen KD & Gade J A simple format for exchange of digitized polygraphic recordings. Electroencephalography and Clinical Neurophysiology 82, 391–393 (1992). [DOI] [PubMed] [Google Scholar]
- [55].Holdgraf C et al. iEEG-BIDS, extending the Brain Imaging Data Structure specification to human intracranial electrophysiology. Scientific data 6, 102 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- [56].Gorgolewski KJ et al. The brain imaging data structure, a format for organizing and describing outputs of neuroimaging experiments. Scientific Data 3, 160044 (2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
- [57].Gramfort A et al. MEG and EEG data analysis with MNE-Python. Frontiers in Neuroscience 7, 267 (2013). [DOI] [PMC free article] [PubMed] [Google Scholar]
- [58].Gramfort A et al. MNE software for processing MEG and EEG data. NeuroImage 86, 446–460 (2014). [DOI] [PMC free article] [PubMed] [Google Scholar]
- [59].Appelhoff S et al. MNE-BIDS: MNE-Python + BIDS = easy dataset interaction. In Organization for Human Brain Mapping (2020). [Google Scholar]
- [60].Van Der Walt S, Colbert SC & Varoquaux G The NumPy array: A structure for efficient numerical computation. Computing in Science and Engineering 13, 22–30 (2011). [Google Scholar]
- [61].Virtanen P et al. SciPy 1.0: fundamental algorithms for scientific computing in Python. Nature Methods 17, 261–272 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- [62].Hunter JD Matplotlib: A 2D graphics environment. Computing in Science and Engineering 9, 99–104 (2007). [Google Scholar]
- [63].Waskom M seaborn: statistical data visualization Journal of Open Source Software (2021). [Google Scholar]
- [64].Allen M, Poggiali D, Whitaker K, Marshall TR & Kievit RA Raincloud plots: A multi-platform tool for robust data visualization [version 1; peer review: 2 approved]. Wellcome Open Research 4, 63 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- [65].Ho J, Tumkaya T, Aryal S, Choi H & Claridge-Chang A Moving beyond P values: data analysis with estimation graphics (2019). [DOI] [PubMed] [Google Scholar]
- [66].Vallat R Pingouin: statistics in Python. Journal of Open Source Software 3, 1026 (2018). [Google Scholar]
- [67].Raschka S MLxtend: Providing machine learning and data science utilities and extensions to Python’s scientific computing stack. Journal of Open Source Software 3, 638 (2018). [Google Scholar]
- [68].Pedregosa F et al. Scikit-learn: Machine Learning in Python. Journal of Machine Learning Research 12, 2825–2830 (2011). [Google Scholar]
- [69].Mezić I Spectral properties of dynamical systems, model reduction and decompositions. Nonlinear Dynamics 41, 309–325 (2005). [Google Scholar]
- [70].Schmid P Dynamic mode decomposition of numerical and experimental data. J. Fluid Mech 656, 5–28 (2010). [Google Scholar]
- [71].Breiman L Random Forests. Machine Learning 45, 5–32 (2001). [Google Scholar]
- [72].Perry R et al. Manifold Forests: Closing the Gap on Neural Networks. Preprint at http://arxiv.org/abs/1909.11799 (2019). [Google Scholar]
- [73].Tomita TM et al. Sparse projection oblique randomer forests. Journal of Machine Learning Research 21 (2020). [Google Scholar]
- [74].Skocik M, Collins J, Callahan-Flintoft C, Bowman H & Wyble B I Tried a Bunch of Things: The Dangers of Unexpected Overfitting in Classification. Preprint at https://www.biorxiv.org/content/10.1101/078816v1.full078816 (2016). [DOI] [PubMed] [Google Scholar]
- [75].Alyakin A, Qin Y & Priebe CE LqRT: Robust Hypothesis Testing of Location Parameters using Lq-Likelihood-Ratio-Type Test in Python. Preprint at http://arxiv.org/abs/1911.11922 (2019). [Google Scholar]
- [76].Qin Y & Priebe CE Robust hypothesis testing via Lq-likelihood. Statistica Sinica 27, 1793–1813 (2017). [Google Scholar]
- [77].Li A, et al. Neural Fragility as an EEG Marker of the Seizure Onset Zone. Gigantum, doiidentifier, 2021. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
We also released the raw iEEG data for patients from NIH, UMH, UMMC, and JHH in the OpenNeuro repository in the form of BIDS-iEEG (https://openneuro.org/datasets/ds003029). Due to restrictions on data sharing from CClinic, we were unable to release the raw iEEG data that we received from this site. Dataset from CClinic is available upon request from authors at the CClinic.