Abstract
Purpose
Accurate prediction of delayed cerebral ischemia (DCI) after subarachnoid hemorrhage (SAH) can be critical for planning interventions to prevent poor neurological outcome. This paper presents a model using convolution dictionary learning to extract features from physiological data available from bedside monitors. We develop and validate a prediction model for DCI after SAH, demonstrating improved precision over standard methods alone.
Methods
488 consecutive SAH admissions from 2006 to 2014 to a tertiary care hospital were included. Models were trained on 80%, while 20% were set aside for validation testing. Modified Fisher Scale was considered the standard grading scale in clinical use; baseline features also analyzed included age, sex, Hunt–Hess, and Glasgow Coma Scales. An unsupervised approach using convolution dictionary learning was used to extract features from physiological time series (systolic blood pressure and diastolic blood pressure, heart rate, respiratory rate, and oxygen saturation). Classifiers (partial least squares and linear and kernel support vector machines) were trained on feature subsets of the derivation dataset. Models were applied to the validation dataset.
Results
The performances of the best classifiers on the validation dataset are reported by feature subset. Standard grading scale (mFS): AUC 0.54. Combined demographics and grading scales (baseline features): AUC 0.63. Kernel derived physiologic features: AUC 0.66. Combined baseline and physiologic features with redundant feature reduction: AUC 0.71 on derivation dataset and 0.78 on validation dataset.
Conclusion
Current DCI prediction tools rely on admission imaging and are advantageously simple to employ. However, using an agnostic and computationally inexpensive learning approach for high-frequency physiologic time series data, we demonstrated that we could incorporate individual physiologic data to achieve higher classification accuracy.
Keywords: subarachnoid hemorrhage, convolutional dictionary learning, time series, machine learning, critical care
Introduction
Subarachnoid hemorrhage (SAH) is a major public health burden, affecting 14.5 per 100,000 persons in the United States alone (1, 2). Much of the resulting functional and cognitive disability is due to delayed cerebral ischemia (DCI) from vasospasm (VSP) (3–7). VSP refers to the narrowing of cerebral blood vessels triggered by the unusual presence of blood surrounding the vessel after a ruptured aneurysm, which can result in stroke. It occurs in 30% of SAH patients (8, 9) [54% of SAH patients in coma (10)]. DCI is a consensus definition with significance for normalizing research efforts in this disease and is defined as the development of new focal neurological signs or decrease of >2 points on the Glasgow Coma Scale (GCS), lasting for more than 1 h, or the appearance of new infarctions on CT or MRI (11, 12), excluding causes other than VSP.
As in other causes of stroke and secondary brain injury in the neurologic intensive care unit (NICU), time is of the essence to detect and intervene. Our interest is in predicting DCI and VSP with greater precision than standard of care scales that rely on admission assessments of blood patterns on computed tomography scans (13–16). For the higher risk SAH patients, the first 10–14 days are occupied by efforts to detect subtle examination changes that suggest VSP (17), and arrange urgent imaging to confirm VSP. For a syndrome with subtle symptoms and time sensitivity, it would be helpful to be more accurate in prediction so clinicians can focus resources, appropriately increase monitoring intensity, and justify diagnostic interventions to prevent permanent injury. On the converse, discharging patients from the ICU at low risk for DCI can result in significant cost savings (18).
Existing predictive models of DCI and VSP after spontaneous SAH are non-dynamic and while they help risk-stratify patients, they can lack accuracy and precision when applied to individuals (13–16). Efforts to improve this early prediction without additional monitoring have met moderate results, by combining risk scores (19), incorporating baseline features such as clinical condition and age (20), or assessment of autoregulation (21).
There is an abundance of physiologic and clinical data that are created and collected in the NICU. Few efforts have explored physiological data for the early prediction of DCI. In Ref. (22), a Naïve Bayes classifier using electronic medical record (EMR) data (cerebrospinal fluid drainage volume, sodium and glucose) and physiologic data [mean arterial blood pressure, heart rate (HR), and intracranial pressure] was able to classify patients for angiographic VSP with a moderately favorable AUC of 0.71. The raw data used in that study was low frequency (hourly at best) and extracted features summarized over 24 or 48 h. Despite the small sample size in that study, the result was encouraging that EMR and physiologic data could improve risk stratification for future events. The question remains whether increased precision can be achieved with use of higher frequency data. In this work, we applied machine-learning techniques, to extract features from the high-frequency data, to predict DCI. There is an extensive literature regarding robust feature extraction from physiological time series data for outcome prediction. Approaches can be broadly classified as either hypothesis driven or data driven. Hypothesis driven approaches have focused primarily on temporal data abstraction that relies on knowledge-based symbolic representations of clinical states, either by a priori threshold setting or interval changes (23, 24), summary statistics (22, 25–28), or template matching (29). Hypothesis driven feature extraction can be effective in prediction but requires domain expertise in designing meta-features and may introduce a bias (25).
Data driven or learning approaches such as used in this study extract meaningful features directly from the labeled data without a priori hypothesis (26, 27, 30–37). Sparse coding and dictionary learning methods (38–42) have shown promise in the field of image processing (38–40, 43–45) and have recently been applied to temporal data (46). Bahadori et al. (36) have used sparse clustering to extract the latent subspace for mortality prediction in the publicly available physionet ICU dataset. Lasko et al. (47) have extracted the temporal dynamics using auto encoders to identify the unlabeled phenotypes expressed in the sequences of serum uric acid signatures of gout vs acute leukemia. This work focuses on sparsity based data-driven techniques to extract the features to remain agnostic about scales, trends or patterns that might be available in the data as opposed to the hypothesis driven which summarizes the temporal data to extract features. In particular, we learned multiscale dictionaries from high-frequency temporal physiologic data that extract informative kernels that maximally classified for DCI.
Materials and Methods
The proposed method is based on recent advances in convolution dictionary learning methods (38–40, 48). Convolution dictionary learning extract translation invariant kernels directly from the data, thus capturing the temporal characteristics of physiological variables. We extracted features using dictionaries learned from patients’ time series data acquired from bedside monitors. We learned multiscale dictionaries using convolution dictionary learning, by down sampling the data at increasing intervals (1, 5, 10, 20, 60, 120, and 240 min). This was intended to capture the temporal dynamics at different resolutions that might be available in the time-series data without a priori hypothesis. We learned the dictionaries for different physiological data variables using convolution dictionary learning as explained in the following sections.
Data analysis and model building were performed using custom software developed in Matlab 2016a (Mathworks, Natick, MA, USA) and Python (www.python.org). All computations were performed using an Intel Xeon CPU 2.2 GHz processor.
Study Population
Consecutive patients with SAH admitted to the NICU between August 1996 and December 2014 were prospectively enrolled in an observational cohort study of SAH patients designed to identify novel risk factors for secondary injury and poor outcome. The study was approved by the Columbia University Medical Center Institutional Review Board. In all cases, written informed consent was obtained from the patient or a surrogate. SAH secondary to perimesencephalic bleeds, trauma, arteriovenous malformation, and patients <18 years old were not enrolled in the study. Starting in 2006, physiologic data was acquired using a high-resolution acquisition system (BedmasterEX; Excel Medical Electronics Inc., Jupiter, FL, USA) from General Electric Solar 8000i monitors (Port Washington, NY, USA; 2006–2013) or Philips Intellivue MP70 monitors (Amsterdam, The Netherlands; 2013–2014) at 0.2 Hz.
Exclusion criteria for this project were the following: (1) absence of physiologic monitoring data (before 2006), (2) VSP or DCI before post bleed day (PBD) 3, and (3) patients missing all candidate features. The targeted classification outcome was DCI, defined as development of new focal neurologic signs or deterioration of consciousness for >1 h or appearance of new infarctions on imaging due to VSP (12).
Baseline Candidate Features
The following baseline characteristics and grading scales were prospectively recorded at admission: age, sex, worst Hunt–Hess grade in first 24 h (HH), mFS, and admission GCS. HH grade was dichotomized into low grade (1–3) and high grade (4–5). MFS was dichotomized into low grade (0–2) and high grade (3–4). Baseline features were compared for patients with DCI vs no DCI. Baseline features were also compared for the derivation vs validation dataset.
Frequency comparisons for categorical variables were performed by Fisher exact test. Two-group comparisons of continuous variables were performed with the Mann–Whitney U test. All statistical tests were two-tailed, and a p-value <0.05 was considered statistically significant.
Physiological Data Extraction
Physiologic data was limited to the first 4 days after aneurysm rupture to limit the influence of clinical treatment in response to suspected VSP or DCI (17). While 0.2 Hz physiological data was available, we remained agnostic about the optimal scale or sampling rate for DCI classification. Five universally available ICU variables [HR, respiratory rate (RR), systolic blood pressure (SBP), diastolic blood pressure (DBP), and oxygen saturation (SPO2)] were downsampled (ds) from 0.2 Hz to 1, 5, 10, 20, 60, 120, and 240 min. Downsampling was computed as means, which also deals with erroneous or missing data (49). These variables were then used to learn the distinct temporal dynamics to derive features.
Feature Extraction Using Convolution Dictionary Learning
Given a time series X ∈ R1×t, our method learns the distinct temporal dynamics by a dictionary based model for each of the five ICU variables. The term “dictionary” refers to the set of basis vectors that can be combined linearly to represent X. These basis vectors are learned directly from the data, and the size of the kernel must be big enough to capture the hidden patterns. In our work, we selected the kernels sizes 2, 5, 10, and 20 to learn multiscale dictionary as explained below.
Dictionary Learning Algorithm
Let X ∈ R1×t be the time series data. For a traditional dictionary learning algorithm, the time series data are then represented by set of patches {xi}i=1,…,N, where xi ∈ R1×r, which may be overlapping patches of size r. Given a set of patches, we learn an over-complete dictionary, denoted by D = {d1, …, dk} ∈ Rr×K where K is the number of basis elements in dictionary D, usually referred to as “atoms” or “kernels.” The patch data are then approximated by DГ, where Г = {γ1, …, γN} is a matrix of sparse vectors. The traditional dictionary learning algorithm then solves the following optimization:
The term denotes the signal reconstruction error. The reconstruction error is minimized subjected to L1 sparsity constraint on sparse vector γi, where λ controls the sparsity. However, using a patch based dictionary results in redundant elements (12). Therefore, we chose to learn the convolution dictionary, which offers two main advantages, (i) the direct support for multiscale dictionaries and (ii) the patch size can be arbitrarily increased at negligible computation cost. The convolution dictionary algorithm learns kernels from the entire time series data instead of patches, thus resulting in fewer atoms. The convolution dictionary solves the following optimization:
where Γk is the sparse activation map over the entire time series data and “*” denotes the convolution operation. By learning the convolution dictionary, we remove many redundant atoms that were simply shifted and clipped versions of the patches (38–40, 50).
To deal with missing/lost data beyond the scale of downsampling, we used mask decoupling (38–40) which introduces a masking operator to zero out missing data while minimizing the error. This requires solving the following optimization:
where W is a mask operator that zeros out any region with the missing data. The above optimization problem is solved using alternating direction method multipliers as described in Boyd et al. (51). Any other convolution dictionary learning method can be used to learn the atoms (52, 53). The dictionaries D, learned at different down sampling rates, are then used for deriving the features for DCI classification.
Feature Computation
Given a time series X ∈ R1×t, dictionary atom γ ∈ R1×k, where k ≤ t, the feature fi for series X and filter γ is given by max(X*γ), where * denotes the valid convolution. Valid convolution means that γ is applied only at each position of X such that γ lies within X. In other words, we performed convolutions only when the contiguous data length was twice the length of kernel. The number of kernels that we extracted from the data was discovered through an optimization step to find maximal model performance. We therefore extracted 20 kernels for each varying kernel length (KL; 2, 5, 8, 10, 20, and 40) and for each downsampling period (ds; 1, 5, 10, 20, 60, 120, and 240 min), and for each of five variables (var; HR, RR, SBP, DBP, and SPO2). This resulted in 4,200 candidate kernel derived physiological features.
Feature Selection and Model Building
We used the above features to develop models using three different classifiers (partial least square, SVM linear, and SVM kernel). Minimal redundancy maximal relevance (mRMR) (54) was applied to identify the most relevant features for classification. mRMR selects the features that maximize the mutual information between features and target class, and minimizes mutual information among the features. The features are ranked based on the greedy search that maximizes the Mutual Information Difference Criterion or Mutual Information Quotient Criterion. Let S ∈ (x1, …, xn) be the set of features and h be the target class (in our case DCI vs non-DCI), then the features are ranked as follows:
where, I(xi, h) is the information gain between the feature xi and target class h. The first “k” ranked features are then used to learn the classifier. This simplifies the model, reduces training times, and enhances the generalizability of the classification model.
We used mRMR in combination with linear and kernel based support vector machines (SVM-L and SMV-K) classifiers (55, 56), as well as partial least squares (PLS) regression (57) for combined feature selection and classification. PLS regression performs a principal component analysis on all feature vectors first and then applies a least squares regression using those components that explain the most variance. Weighted SVM (58) was utilized to account for the imbalance in classification categories (i.e., fewer DCI vs non-DCI in any consecutive SAH dataset). We created models using baseline features, physiological features and combined baseline and physiological features, along with feature selection to test the discriminative ability of different features. We compared the performance of the physiological features learned from our model with the baseline and grading scale features.
Internal Validation and Validation Strategy
The cohort was randomly split 80/20%, while maintaining proportional targeted outcome (DCI). 80% were used to learn the dictionary and train/test models, and considered the primary derivation dataset. For internal validation of our models, we performed cross-validation of the derivation data with a 12.5% hold-out set; the hold-out set was proportional to the training data set for percentage of targeted outcome. Discriminative performance is described by an area under the receiver operating characteristic curve (AUC), and AUCs were statistically compared (59). The median value of AUC is reported, over 100 runs. 20% of the cohort were not involved in model training, and used exclusively for testing the classification accuracy of our models. Classification accuracy of our models on the validation test set is reported as AUC, with 95% confidence intervals (CI). An overview of the analytical approach is illustrated in Figure 1. To summarize, physiological variables were downsampled to extract temporal patterns at varying scales using dictionary learning. We extracted 20 kernels for each varying KL (2, 5, 8, 10, 20, and 40) and for each downsampling period (ds; 1, 5, 10, 20, 60, 120, and 240 min), and for each of five variables (var; HR, RR, SBP, DBP, and SPO2). The resulting kernels were convolved with time series at different scales to extract the maximal value resulting in 4,200 features; the dimensions of these features were reduced by mRMR to classify DCI using PLS, SVM-L, and SVM-K.
Results
From May 2006 to December 2014, 562 SAH patients with physiologic data were enrolled. 8 had VSP or DCI identified before PBD 3, 66 were missing all candidate features leaving a total of 488 subjects included in the study. The median AUC of 100 runs of cross-validation (with 12.5% hold-out set) is presented in Table 1.
Table 1.
Features | Derivation dataset (median AUC of 100 runs) |
Validation dataset [AUC (95% confidence intervals)] |
||||
---|---|---|---|---|---|---|
Classifiers |
Classifiers |
|||||
PLS | SVM-L | SVM-K | PLS | SVM-L | SVM-K | |
Age | 0.58 | 0.54 | 0.53 | 0.58 (0.46–0.7) | 0.6 (0.48–0.71) | 0.64 (0.53–0.76) |
Sex | 0.59 | 0.59 | 0.59 | 0.62 (0.5–0.74) | 0.62 (0.5–0.74) | 0.62 (0.5–0.74) |
Hunt Hess Scale | 0.60 | 0.55 | 0.58 | 0.49 (0.37–0.61) | 0.46 (0.34–0.58) | 0.5 (0.38–0.62) |
Modified Fisher Scale | 0.54 | 0.57 | 0.50 | 0.47 (0.35–0.59) | 0.53 (0.41–0.65) | 0.53 (0.41–0.65) |
Glasgow Coma Scale | 0.59 | 0.57 | 0.63 | 0.43 (0.31–0.55) | 0.44 (0.32–0.56) | 0.56 (0.44–0.68) |
Baseline (age, sex, and scales) | 0.63 | 0.58 | 0.54 | 0.64 (0.53–0.76) | 0.59 (0.48–0.71) | 0.61 (0.49–0.72) |
Diastolic blood pressure | 0.52 | 0.48 | 0.56 | 0.44 (0.26–0.61) | 0.42 (0.25–0.59) | 0.56 (0.19–0.53) |
Systolic blood pressure | 0.58 | 0.54 | 0.49 | 0.65 (0.49–0.82) | 0.43 (0.26–0.6) | 0.36 (0.19–0.53) |
Heart rate | 0.55 | 0.51 | 0.50 | 0.46 (0.28–0.63) | 0.5 (0.33–0.68) | 0.45 (0.28–0.62) |
Oxygen saturation | 0.56 | 0.53 | 0.50 | 0.62 (0.45–0.79) | 0.48 (0.31–0.65) | 0.5 (0.33–0.67) |
Respiratory rate | 0.49 | 0.50 | 0.50 | 0.57 (0.4–0.74) | 0.54 (0.36–0.71) | 0.5 (0.33–0.67) |
Combined physiological | 0.66 | 0.56 | 0.50 | 0.47 (0.3–0.64) | 0.51 (0.34–0.68) | 0.5 (0.33–0.67) |
Baseline and physiological | 0.63 | 0.56 | 0.50 | 0.5 (0.33–0.67) | 0.5 (0.33–0.67) | 0.5 (0.33–0.67) |
MRMR (baseline and physiological) | 0.71 | 0.60 | 0.50 | 0.78 (0.64–0.92) | 0.64 (0.47–0.8) | 0.5 (0.33–0.67) |
The SVM-L classifier with maximal relevance and minimal redundancy (MRMR) feature reduction performed the best.
Values highlighted in bold indicates the performance of the classifier that performed the best.
Baseline Feature Model Performance
Among demographical information, sex (AUC 0.59) performed slightly better than age (AUC 0.58, PLS). GCS (AUC 0.63, SVM-K) achieved slightly better accuracy than HH (AUC 0.60, PLS) and MFS (AUC 0.57, SVM-L). By combining demographics and grading scales (age, sex, HH, mFS, and GCS), a PLS classifier performed better than the individual features with an AUC of 0.63 in predicting DCI.
The classification accuracy on the validation set was found to be similar to the derivation set. A model based on current standard grading scale (MFS) achieved an AUC of 0.56 (SVM-L, 95% CI, 0.44–0.67). Combining all the demographics and grading scales improved the AUC to 0.64 (PLS, 95% CI, 0.52–0.76).
Physiological Feature Model Performance
Features extracted from individual physiological time-series variables did not perform significantly better than the baseline features. SBP (AUC 0.58, PLS) achieved slightly better AUC than the other variables. However, adding all the features derived from physiological data achieved an AUC of 0.66. Adding demographics and grading scales along with the feature reduction performed better than individual features by achieving an AUC of 0.71, which was statistically significantly higher than the performance of MFS (AUC of 0.57, p = 0.0025).
The performance on the validation set was found to be similar to that of the derivation set. Feature reduction (to reduce redundancy and maximal relevance) when applied to combined demographics, grading scales, and physiological data produced the best classification performance with an AUC of 0.78 (PLS, 95% CI, 0.63–0.92).
In the case of the PLS classifier, the weights indicate the discriminative power of the features in separating the two classes. Figure 2 shows the PLS weights of the 80 features. Figure 3 shows the kernels corresponding to a demonstrative selection of the top 10 features relevant for classification. The kernel displays the time varying characteristics for different variables and highlights the need for capturing high-frequency data at different scales (downsampling rate).
Discussion
Recognizing trends and patterns, and minutely analyzing complex data requires the layered knowledge of clinical experts, but defies rule-based systems. It has previously been shown that a Naïve Bayes classifier using summary statistics of 24 h (low frequency) data can classify for angiographic VSP better than clinician-dependent Dopplers and exams (22). Here, we show that features extracted from higher frequency temporal data (ranging from 1 min to 4 h) is superior to gold standard grading scales in classification of DCI after SAH. In our approach, we extracted high-level features from existing physiologic data, without an a priori hypothesis of what patterns might emerge. There are two parts to our algorithm; first is extracting kernels from the data using convolution dictionary learning, which is computationally expensive but can be performed offline and once (kernel generation for five variables for six KLs and for seven downsampling rates took 1 day). Second is the model building, this is relatively faster and is on the order of minutes; this, too, can be performed offline and once. Once the model is built, it can be applied for each patient in a clinical setting on day 4 of SAH, with a computational time on the order of seconds. Our approach is more accurate than a gold standard grading scale and viable in a clinical setting.
To enable validation efforts and generalizability to other datasets and institutions, we focused on universal physiologic ICU variables and typical baseline grading scales pertinent to SAH used in the NICU. In this translational work, we used a dictionary learning method to extract frequency selective and translation invariant characteristics of time series data. To the best of our knowledge, this is the first study that shows the efficacy of dictionary learning for DCI classification using time series. The novel application to time series data required some choices bound by characteristics of the dataset (KLs) and domain (downsampling rates). We tested our method for its discriminative ability for DCI and found that dictionary kernel derived physiological features outperformed a gold standard static grading scale. When combined with grading scales and demographics, our dictionary learning based method predicted DCI with an AUC (0.71, PLS) approaching clinical reliability [threshold of 0.8 (60)]. A model with MFS alone had an AUC of 0.57. While a comparison of AUCs is frequently used in biostatistics and computer science to demonstrate trends of improvement between tools, it is not sensitive enough to accurately capture improvements in predictive discrimination (61). We therefore compared the AUCs using Hanley’s method of comparing ROCs of specific tools in the same population, which showed the statistical significance of the difference between the two models’ ROCs (59).
An effort to show robustness of the model was with an internal validation strategy, testing on a separate dataset excluded from model building entirely. Generalizability of a machine-learning algorithm, however, assumes that the training dataset is large and diverse enough to be representative. A limitation of our study is the possibility of causality leakage (62). We attempted to limit effects of causality leakage, i.e., the influence of cerebral perfusion efforts on our data, by censoring beyond day 4, which is the highest risk of onset of DCI. Another limitation to this study is the single center approach; there is no publicly available dataset for SAH with similar granularity of physiologic data. Future efforts will include developing complementary SAH cohorts and validating these algorithms on other centers’ data.
Conclusion
A data-driven dictionary based featurization and learning approach to physiological time series data prior to peak DCI period shows promise to improve prediction precision. This is a computationally inexpensive and agnostic feature extraction approach for physiologic time series parameters in the ICU (HR, RR, SBP, DBP, and SPO2). There is a vast pool of candidate features within the EMR with a biological basis for classification ability (i.e., drawn from frequentist statistical studies showing relationship with VSP and DCI in specific SAH cohorts). Future efforts will also draw from this feature pool to further improve the precision of DCI prediction, favoring those candidate features that are obtained for standard clinical care and thus potentially automatable.
Ethics Statement
Consecutive patients with aneurysmal SAH admitted to the NICU between August 1996 and December 2014 were prospectively enrolled in an observational cohort study of SAH patients designed to identify novel risk factors for secondary injury and poor outcome. The study was approved by the Columbia University Medical Center Institutional Review Board. In all cases, written informed consent was obtained from the patient or a surrogate.
Author Contributions
SP, JC, SA, DR, EC, AV, KD and HF: Data collection. MM, SP, and KT: Analysis. MM and SP: writing. MM, KT, HF, AV, KD, EC, DR, SA, JC, NE and SP: Editing.
Conflict of Interest Statement
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest. The reviewers LC and IS and handling editor declared their shared affiliation.
Acknowledgments
The study was supported by National Institutes of Health K01-ES026833-02 (SP) and National Science Foundation 1344668 (NE). The authors wish to thank Brendt Wohlberg for the availability of dictionary learning software (https://github.com/bwohlberg/sporco) and for his valuable discussions on convolution dictionary learning.
References
- 1.Shea AM, Reed SD, Curtis LH, Alexander MJ, Villani JJ, Schulman KA. Characteristics of nontraumatic subarachnoid hemorrhage in the United States in 2003. Neurosurgery (2007) 61(6):1131–7. 10.1227/01.neu.0000306090.30517.ae [DOI] [PubMed] [Google Scholar]
- 2.Qureshi AI, Suri MF, Nasar A, Kirmani JF, Divani AA, He W, et al. Trends in hospitalization and mortality for subarachnoid hemorrhage and unruptured aneurysms in the United States. Neurosurgery (2005) 57(1):1–8. 10.1227/01.NEU.0000163081.55025.CD [DOI] [PubMed] [Google Scholar]
- 3.Roos YB, Dijkgraaf MG, Albrecht KW, Beenen LF, Groen RJ, de Haan RJ, et al. Direct costs of modern treatment of aneurysmal subarachnoid hemorrhage in the first year after diagnosis. Stroke (2002) 33(6):1595–9. 10.1161/01.STR.0000016401.49688.2F [DOI] [PubMed] [Google Scholar]
- 4.Mayer SA, Kreiter KT, Copeland D, Bernardini GL, Bates JE, Peery S, et al. Global and domain-specific cognitive impairment and outcome after subarachnoid hemorrhage. Neurology (2002) 59(11):1750–8. 10.1212/01.WNL.0000035748.91128.C2 [DOI] [PubMed] [Google Scholar]
- 5.Hackett ML, Anderson CS. Health outcomes 1 year after subarachnoid hemorrhage: an international population-based study. The Australian cooperative research on subarachnoid hemorrhage study group. Neurology (2000) 55(5):658–62. 10.1212/WNL.55.5.658 [DOI] [PubMed] [Google Scholar]
- 6.Charpentier C, Audibert G, Guillemin F, Civit T, Ducrocq X, Bracard S, et al. Multivariate analysis of predictors of cerebral vasospasm occurrence after aneurysmal subarachnoid hemorrhage. Stroke (1999) 30(7):1402–8. 10.1161/01.STR.30.7.1402 [DOI] [PubMed] [Google Scholar]
- 7.Dorsch N. A clinical review of cerebral vasospasm and delayed ischaemia following aneurysm rupture. Acta Neurochir Suppl (2011) 110(Pt 1):5–6. 10.1007/978-3-7091-0353-1_1 [DOI] [PubMed] [Google Scholar]
- 8.Schmidt JM, Wartenberg KE, Fernandez A, Claassen J, Rincon F, Ostapkovich ND, et al. Frequency and clinical impact of asymptomatic cerebral infarction due to vasospasm after subarachnoid hemorrhage. J Neurosurg (2008) 109(6):1052–9. 10.3171/JNS.2008.109.12.1052 [DOI] [PubMed] [Google Scholar]
- 9.Rabinstein AA, Pichelmann MA, Friedman JA, Piepgras DG, Nichols DA, McIver JI, et al. Symptomatic vasospasm and outcomes following aneurysmal subarachnoid hemorrhage: a comparison between surgical repair and endovascular coil occlusion. J Neurosurg (2003) 98(2):319–25. 10.3171/jns.2003.98.2.0319 [DOI] [PubMed] [Google Scholar]
- 10.Kirmani JF, Qureshi AI, Hanel RA, Siddiqui AM, Safdar A, Yahia AM, et al. Silent cerebral infarctions in poor-grade patients with subarachnoid hemorrhage. Neurology (2002) 58(7):A159–159. [Google Scholar]
- 11.Frontera JA, Fernandez A, Schmidt JM, Claassen J, Wartenberg KE, Badjatia N, et al. Defining vasospasm after subarachnoid hemorrhage: what is the most clinically relevant definition? Stroke (2009) 40(6):1963–8. 10.1161/STROKEAHA.108.544700 [DOI] [PubMed] [Google Scholar]
- 12.Vergouwen MD, Vermeulen M, van Gijn J, Rinkel GJ, Wijdicks EF, Muizelaar JP, et al. Definition of delayed cerebral ischemia after aneurysmal subarachnoid hemorrhage as an outcome event in clinical trials and observational studies: proposal of a multidisciplinary research group. Stroke (2010) 41(10):2391–5. 10.1161/STROKEAHA.110.589275 [DOI] [PubMed] [Google Scholar]
- 13.Rosen DS, Macdonald RL. Subarachnoid hemorrhage grading scales: a systematic review. Neurocrit Care (2005) 2(2):110–8. 10.1385/NCC:2:2:110 [DOI] [PubMed] [Google Scholar]
- 14.Fisher CM, Kistler JP, Davis JM. Relation of cerebral vasospasm to subarachnoid hemorrhage visualized by computerized tomographic scanning. Neurosurgery (1980) 6(1):1–9. 10.1097/00006123-198001000-00001 [DOI] [PubMed] [Google Scholar]
- 15.Claassen J, Bernardini GL, Kreiter K, Bates J, Du YE, Copeland D, et al. Effect of cisternal and ventricular blood on risk of delayed cerebral ischemia after subarachnoid hemorrhage: the Fisher scale revisited. Stroke (2001) 32(9):2012–20. 10.1161/hs0901.095677 [DOI] [PubMed] [Google Scholar]
- 16.Frontera JA, Claassen J, Schmidt JM, Wartenberg KE, Temes R, Connolly ES, Jr, et al. Prediction of symptomatic vasospasm after subarachnoid hemorrhage: the modified Fisher scale. Neurosurgery (2006) 59(1):21–7. 10.1227/01.NEU.0000218821.34014.1B [DOI] [PubMed] [Google Scholar]
- 17.Heros RC, Zervas NT, Varsos V. Cerebral vasospasm after subarachnoid hemorrhage: an update. Ann Neurol (1983) 14(6):599–608. 10.1002/ana.410140602 [DOI] [PubMed] [Google Scholar]
- 18.Crobeddu E, Mittal MK, Dupont S, Wijdicks EF, Lanzino G, Rabinstein AA. Predicting the lack of development of delayed cerebral ischemia after aneurysmal subarachnoid hemorrhage. Stroke (2012) 43(3):697–701. 10.1161/STROKEAHA.111.638403 [DOI] [PubMed] [Google Scholar]
- 19.Foreman PM, Chua MH, Harrigan MR, Fisher WS, III, Tubbs RS, Shoja MM, et al. External validation of the practical risk chart for the prediction of delayed cerebral ischemia following aneurysmal subarachnoid hemorrhage. J Neurosurg (2016) 126(5):1530–6. 10.3171/2016.1.JNS152554 [DOI] [PubMed] [Google Scholar]
- 20.de Rooij NK, Greving JP, Rinkel GJ, Frijns CJ. Early prediction of delayed cerebral ischemia after subarachnoid hemorrhage: development and validation of a practical risk chart. Stroke (2013) 44(5):1288–94. 10.1161/STROKEAHA.113.001125 [DOI] [PubMed] [Google Scholar]
- 21.Calviere L, Nasr N, Arnaud C, Czosnyka M, Viguier A, Tissot B, et al. Prediction of delayed cerebral ischemia after subarachnoid hemorrhage using cerebral blood flow velocities and cerebral autoregulation assessment. Neurocrit Care (2015) 23(2):253–8. 10.1007/s12028-015-0125-x [DOI] [PubMed] [Google Scholar]
- 22.Roederer A, Holmes JH, Smith MJ, Lee I, Park S. Prediction of significant vasospasm in aneurysmal subarachnoid hemorrhage using automated data. Neurocrit Care (2014) 21(3):444–50. 10.1007/s12028-014-9976-9 [DOI] [PubMed] [Google Scholar]
- 23.Sacchi L, Dagliati A, Bellazzi R. Analyzing complex patients’ temporal histories: new frontiers in temporal data mining. Data Mining Clin Med (2015) 1246:89–105. 10.1007/978-1-4939-1985-7_6 [DOI] [PubMed] [Google Scholar]
- 24.Stacey M, McGregor C. Temporal abstraction in intelligent clinical data analysis: a survey. Artif Intell Med (2007) 39(1):1–24. 10.1016/j.artmed.2006.08.002 [DOI] [PubMed] [Google Scholar]
- 25.Verduijn M, Sacchi L, Peek N, Bellazzi R, de Jonge E, de Mol BA. Temporal abstraction for feature extraction: a comparative case study in prediction from intensive care monitoring data. Artif Intell Med (2007) 41(1):1–12. 10.1016/j.artmed.2007.06.003 [DOI] [PubMed] [Google Scholar]
- 26.Saria S, Rajani AK, Gould J, Koller D, Penn AA. Integration of early physiological responses predicts later illness severity in preterm infants. Sci Transl Med (2010) 2(48):48ra65. 10.1126/scitranslmed.3001304 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Saria S, Koller D, Penn A. Learning individual and population level traits from clinical temporal data. Proceedings of the Neural Information Processing Systems (NIPS) Predictive Models in Personalized Medicine Whistler (2010). [Google Scholar]
- 28.Mayer CC, Bachler M, Hortenhuber M, Stocker C, Holzinger A, Wassertheurer S. Selection of entropy-measure parameters for knowledge discovery in heart rate variability data. BMC Bioinform (2014) 15(Suppl 6):S2. 10.1186/1471-2105-15-S6-S2 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Dua S, Saini S, Singh H. Temporal pattern mining for multivariate time series classification. J Med Imag Health Inform (2011) 1(2):164–9. 10.1166/jmihi.2011.1019 [DOI] [Google Scholar]
- 30.Lehman LW, Adams RP, Mayaud L, Moody GB, Malhotra A, Mark RG, et al. A physiological time series dynamics-based approach to patient monitoring and outcome prediction. IEEE J Biomed Health Inform (2015) 19(3):1068–76. 10.1109/JBHI.2014.2330827 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Schulam P, Wigley F, Saria S. Clustering longitudinal clinical marker trajectories from electronic health data: applications to phenotyping and endotype discovery. Proceedings of the AAAI Austin: CiteseerX (2015). p. 2956–64. [Google Scholar]
- 32.Nemati S, Adams R. Supervised learning in dynamic Bayesian networks. Proceedings of the Neural Information Processing Systems (NIPS) Workshop on Deep Learning and Representation Learning Phoenix (2014). [Google Scholar]
- 33.Luo Y, Xin Y, Joshi R, Celi L, Szolovits P. Predicting icu mortality risk by grouping temporal trends from a multivariate panel of physiologic measurements. Proceedings of the AAAI Phoenix (2016). p. 42–50. [Google Scholar]
- 34.Lipton ZC, Kale DC, Wetzell RC. Phenotyping of Clinical Time Series with LSTM Recurrent Neural Networks (2015). https://arxiv.org/abs/1510.07641
- 35.Kale DC, Gong D, Che Z, Liu Y, Medioni G, Wetzel R, et al. An examination of multivariate time series hashing with applications to health care. Proceedings of the 2014 IEEE International Conference on Data Mining (ICDM) Shenzhen: IEEE (2014). p. 260–9. [Google Scholar]
- 36.Bahadori MT, Kale DC, Fan Y, Liu Y. Functional subspace clustering with application to time series. Proceedings of the ICML Lille (2015). p. 228–37. [Google Scholar]
- 37.Marlin BM, Kale DC, Khemani RG, Wetzel RC. Unsupervised pattern discovery in electronic health care data using probabilistic clustering models. Proceedings of the 2nd ACM SIGHIT International Health Informatics Symposium Miami, FL: ACM (2012). p. 389–98. [Google Scholar]
- 38.Wohlberg B. Efficient algorithms for convolutional sparse representations. IEEE Trans Image Process (2016) 25(1):301–15. 10.1109/TIP.2015.2495260 [DOI] [PubMed] [Google Scholar]
- 39.Wohlberg B. Convolutional sparse representations as an image model for impulse noise restoration. Proceedings of the 2016 IEEE 12th Image, Video, and Multidimensional Signal Processing Workshop (IVMSP) Bordeaux: IEEE (2016). p. 1–5. [Google Scholar]
- 40.Wohlberg B. Boundary handling for convolutional sparse representations. Proceedings of the Image Processing (ICIP), 2016 IEEE International Conference on Phoenix: IEEE (2016). p. 1833–7. [Google Scholar]
- 41.Mairal J, Bach F, Ponce J, Sapiro G. Online dictionary learning for sparse coding. Proceedings of the 26th Annual International Conference on Machine Learning Montreal: ACM (2009). p. 689–96. [Google Scholar]
- 42.Aharon M, Elad M, Bruckstein A. SVD: an algorithm for designing overcomplete dictionaries for sparse representation. IEEE Trans Sig Process (2006) 54(11):4311–22. 10.1109/TSP.2006.881199 [DOI] [Google Scholar]
- 43.Elad M, Aharon M. Image denoising via sparse and redundant representations over learned dictionaries. IEEE Trans Image Process (2006) 15(12):3736–45. 10.1109/TIP.2006.881969 [DOI] [PubMed] [Google Scholar]
- 44.Megjhani M, Rey-Villamizar N, Merouane A, Lu Y, Mukherjee A, Trett K, et al. Population-scale three-dimensional reconstruction and quantitative profiling of microglia arbors. Bioinformatics (2015) 31(13):2190–8. 10.1093/bioinformatics/btv109 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Megjhani M, Correa, de Sampaio P, Leigh Carstens J, Kalluri R, Roysam B. Morphologically constrained spectral unmixing by dictionary learning for multiplex fluorescence microscopy. Bioinformatics (2017) 33(14):2182–90. 10.1093/bioinformatics/btx108 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Cogliati A, Duan Z, Wohlberg B. Piano music transcription with fast convolutional sparse coding. Proceedings of the 2015 IEEE 25th International Workshop on Machine Learning for Signal Processing (MLSP) Boston: IEEE (2015). p. 1–6. [Google Scholar]
- 47.Lasko TA, Denny JC, Levy MA. Computational phenotype discovery using unsupervised feature learning over noisy, sparse, and irregular clinical data. PLoS One (2013) 8(6):e66341. 10.1371/journal.pone.0066341 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48.Wohlberg B. Efficient convolutional sparse coding. Proceedings of the 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) Florence: IEEE (2014). p. 7173–7. [Google Scholar]
- 49.Johnson AE, Ghassemi MM, Nemati S, Niehaus KE, Clifton DA, Clifford GD. Machine learning and decision support in critical care. Proc IEEE Inst Electr Electron Eng (2016) 104(2):444–66. 10.1109/JPROC.2015.2501978 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.Kavukcuoglu K, Sermanet P, Boureau Y-L, Gregor K, Mathieu M, Cun YL. Learning convolutional feature hierarchies for visual recognition. Proceedings of the Advances in Neural Information Processing Systems Vancouver (2010). p. 1090–8. [Google Scholar]
- 51.Boyd S, Parikh N, Chu E, Peleato B, Eckstein J. Distributed optimization and statistical learning via the alternating direction method of multipliers. Found Trends Mach Learn (2011) 3(1):1–122. 10.1561/2200000016 [DOI] [Google Scholar]
- 52.Akhtar N, Shafait F, Mian A. Discriminative Bayesian dictionary learning for classification. IEEE Trans Pattern Anal Mach Intell (2016) 38(12):2374–88. 10.1109/TPAMI.2016.2527652 [DOI] [PubMed] [Google Scholar]
- 53.Yang M, Liu W, Luo W, Shen L. Analysis-synthesis dictionary learning for universality-particularity representation based classification. Proceedings of the 30th AAAI Conference on Artificial Intelligence Phoenix (2016). [Google Scholar]
- 54.Peng H, Long F, Ding C. Feature selection based on mutual information: criteria of max-dependency, max-relevance, and min-redundancy. IEEE Trans Pattern Anal Mach Intell (2005) 27(8):1226–38. 10.1109/TPAMI.2005.159 [DOI] [PubMed] [Google Scholar]
- 55.Huang YM, Du SX. Weighted support vector machine for classification with uneven training class sizes. Proceedings of the 2005 International Conference on Machine Learning and Cybernetics, Vols 1-9 Guangzhou (2005). p. 4365–9. [Google Scholar]
- 56.Cortes C, Vapnik V. Support-vector networks. Mach Learn (1995) 20(3):273–97. 10.1007/bf00994018 [DOI] [Google Scholar]
- 57.Geladi P, Kowalski BR. Partial least-squares regression: a tutorial. Anal Chim Acta (1986) 185:1–17. 10.1016/0003-2670(86)80028-9 [DOI] [Google Scholar]
- 58.Chang CC, Lin CJ. LIBSVM: a library for support vector machines. ACM Trans Intell Syst Technol (2011) 2(3):27. 10.1145/1961189.1961199 [DOI] [Google Scholar]
- 59.Hanley JA, McNeil BJ. A method of comparing the areas under receiver operating characteristic curves derived from the same cases. Radiology (1983) 148(3):839–43. 10.1148/radiology.148.3.6878708 [DOI] [PubMed] [Google Scholar]
- 60.Steyerberg EW, Pencina MJ, Lingsma HF, Kattan MW, Vickers AJ, Van Calster B. Assessing the incremental value of diagnostic and prognostic markers: a review and illustration. Eur J Clin Invest (2012) 42(2):216–28. 10.1111/j.1365-2362.2011.02562.x [DOI] [PMC free article] [PubMed] [Google Scholar]
- 61.Lobo JM, Jimenez-Valverde A, Real R. AUC: a misleading measure of the performance of predictive distribution models. Global Ecol Biogeography (2008) 17(2):145–51. 10.1111/j.1466-8238.2007.00358.x [DOI] [Google Scholar]
- 62.Kaufman S, Rosset S, Perlich C, Stitelman O. Leakage in data mining: formulation, detection, and avoidance. ACM Trans Knowl Discov Data (2012) 6(4):15. 10.1145/2382577.2382579 [DOI] [Google Scholar]