Abstract
Background
Diagnostic classification of central vs. peripheral etiologies in acute vestibular disorders remains a challenge in the emergency setting. Novel machine-learning methods may help to support diagnostic decisions. In the current study, we tested the performance of standard and machine-learning approaches in the classification of consecutive patients with acute central or peripheral vestibular disorders.
Methods
40 Patients with vestibular stroke (19 with and 21 without acute vestibular syndrome (AVS), defined by the presence of spontaneous nystagmus) and 68 patients with peripheral AVS due to vestibular neuritis were recruited in the emergency department, in the context of the prospective EMVERT trial (EMergency VERTigo). All patients received a standardized neuro-otological examination including videooculography and posturography in the acute symptomatic stage and an MRI within 7 days after symptom onset. Diagnostic performance of state-of-the-art scores, such as HINTS (Head Impulse, gaze-evoked Nystagmus, Test of Skew) and ABCD2 (Age, Blood, Clinical features, Duration, Diabetes), for the differentiation of vestibular stroke vs. peripheral AVS was compared to various machine-learning approaches: (i) linear logistic regression (LR), (ii) non-linear random forest (RF), (iii) artificial neural network, and (iv) geometric deep learning (Single/MultiGMC). A prospective classification was simulated by ten-fold cross-validation. We analyzed whether machine-estimated feature importances correlate with clinical experience.
Results
Machine-learning methods (e.g., MultiGMC) outperform univariate scores, such as HINTS or ABCD2, for differentiation of all vestibular strokes vs. peripheral AVS (MultiGMC area-under-the-curve (AUC): 0.96 vs. HINTS/ABCD2 AUC: 0.71/0.58). HINTS performed similarly to MultiGMC for vestibular stroke with AVS (AUC: 0.86), but more poorly for vestibular stroke without AVS (AUC: 0.54). Machine-learning models learn to put different weights on particular features, each of which is relevant from a clinical viewpoint. Established non-linear machine-learning methods like RF and linear methods like LR are less powerful classification models (AUC: 0.89 vs. 0.62).
Conclusions
Established clinical scores (such as HINTS) provide a valuable baseline assessment for stroke detection in acute vestibular syndromes. In addition, machine-learning methods may have the potential to increase sensitivity and selectivity in the establishment of a correct diagnosis.
Keywords: Acute vestibular syndrome, HINTS, Machine-learning, MRI, Vestibular neuritis, Vestibular stroke
Introduction
Patients with acute vertigo and dizziness account for about 4% of all visits to the emergency department (ED) [1]. Stroke is the underlying cause in 4–15% of all patients, and up to 25% of patients with the presentation of acute vestibular syndrome (AVS, defined by the presence of spontaneous nystagmus) [1, 2]. About 10% of strokes are missed at first contact [3]. Patients discharged from the ED with a suspected benign diagnosis of acute vertigo or dizziness have a 50-fold increased risk of stroke in the first week compared to matched controls [4]. Reasons for this deplorable situation are an overreliance on symptom quality and intensity as distinctive features, inadequate knowledge or application of bedside ocular motor examinations, and a blind trust in cerebral imaging results [5]. Consequently, ED physicians worldwide rank vertigo and dizziness as one of the top priorities for the development of better diagnostic algorithms [6].
Different concepts exist to differentiate peripheral and central etiologies of acute vertigo and dizziness [7, 8]. One strategy relies on a comprehensive examination of vestibular, ocular motor, and postural functions. For AVS, the HINTS test (Head Impulse, gaze-evoked Nystagmus, Test of Skew) has a high sensitivity and specificity (> 90%) for identification of stroke [9]. The diagnostic accuracy of HINTS can be further improved by video oculographic quantification of the head impulse test (vHIT) [10, 11]. Examination-based classification approaches require a profound knowledge of examination techniques and expertise in interpretation of findings. Another idea is to stratify the risk of vestibular stroke by diagnostic index tests, which aggregate information on symptom characteristics (such as symptom onset and duration, triggers, accompanying complaints) and cardiovascular risk factors (CVRF). For example, the ABCD2 score (Age, Blood pressure, Clinical features, Duration, Diabetes) can help to estimate the risk of vestibular stroke, but is inferior to HINTS in diagnostic accuracy [12, 13]. The advantage of index tests based on history taking is that they are easy to apply and not restricted to clinical subtypes such as AVS. Diagnostic approaches by magnetic resonance imaging (MRI) only, have a high rate of false-negative results (50% for lesions < 10 mm) in the first 48 h after symptom onset and are, therefore, not reliable during the acute stage [5, 14].
In the current study, we applied modern machine-learning algorithms to classify vestibular stroke vs. peripheral AVS due to vestibular neuritis based on a multimodal data set (including a standardized assessment of symptom features, CVRF, and detailed quantitative testing of ocular motor, vestibular, and postural functions). Machine-learning approaches were compared to state-of-the-art tests (such as HINTS, ABCD2) to evaluate their feasibility and value for diagnostic decision support.
Methods
Patient cohorts and study protocol
In total 108 patients, who were admitted to the ED of the University Hospital (LMU Munich), were included in this study and received a standardized assessment (of symptom features, CVRF, and vestibular, ocular motor and postural functions) following the EMVERT trial protocol [15]. Based on the findings of MRI (performed within 7 days after symptom onset) and videooculography (vHIT gain threshold: 0.7, refixation saccades, gaze-evoked nystagmus, skew deviation), 40 patients were diagnosed as having vestibular stroke (64.1 ± 12.2 years, 67.5% men, 19 with presentation of AVS), and 68 as having peripheral AVS due to vestibular neuritis (55.6 ± 14.6 years, 64.7% men). Classification algorithms (established index tests vs. modern machine-learning techniques) were applied post hoc to test their diagnostic accuracy for differentiation of both groups.
Protocol approval and patient consent
The study was approved by the Ethics Committee of the University of Munich on February 23, 2015 (57–15). The study was conducted according to the Guideline for Good Clinical Practice (GCP), the Federal Data Protecting Act and the Helsinki Declaration of the World Medical Association in its current version (revision of Fortaleza, Brazil, October 2013). All subjects gave their informed, written consent to participate in the study.
Assessment of symptom characteristics and cardiovascular risk factors
In all patients, a standardized history was taken in the ED, including the following features: symptom quality (vertigo, dizziness, double vision), symptom onset (acute, lingering), symptom duration (10–60 min, > 60 min), symptom intensity (by visual analogue scale), preceding triggers (yes, no), accompanying features (ear symptoms, central neurological symptoms), and CVRF (diabetes, high blood pressure (> 140 mmHg), nicotine abuse, atrial fibrillation, family history, prior stroke or myocardial infarction). Health-related quality of life and functional impairment was assessed by questionnaires: European Quality of Life Score—5 dimensions—5 levels (EQ-5D-5L), including subscores for anxiety, pain, activity, self-care, and mobility (ranging from 1–5 each with 5 indicating worst impairment) [16], EQ visual analogue scale (EQ-VAS) (ranging from 0–100 with 100 being the best status), Dizziness Handicap Inventory (DHI) (ranging from 0–100 points (maximum)) [17], and modified Rankin scale (mRS) (ranging from 0–6 points).
Quantitative assessment of vestibular, ocular motor and postural functions
Videooculography (VOG): Vestibular and ocular motor signs were documented by VOG (EyeSeeCam®, EyeSeeTec GmbH, Munich, Germany) during the acute stage of symptoms, including nystagmus in straight ahead position (slow phase velocity (SPV) (°/sec), amplitude (°), horizontal and vertical component, with and without fixation), horizontal vestibulo-ocular reflex (VOR) function by vHIT (gain, presence of refixation saccades), gaze-evoked nystagmus (SPV (°/sec), horizontal and vertical component, lateral and vertical gaze positions), saccades (velocity (°/sec), horizontal and vertical direction), smooth pursuit (gain, horizontal and vertical direction), fixation suppression of the VOR (gain, horizontal direction), and skew deviation (cover test in six gaze positions). VOR gain was rated as pathological for values < 0.7. Suppression of spontaneous nystagmus (SPN) was positive, if the horizontal or vertical component of the SPV decreased by at least 40% on fixation.
Testing of subjective visual vertical (SVV): The SVV was measured by the bucket test method as described previously [18, 19]. Ten repetitions (5 clockwise/ 5 counter clockwise rotations) were performed and a mean of the deviations was calculated. The normal range was defined as 0 ± 2.5° [19].
Posturography: A posturographic measurement of body sway was performed using a mobile device (Wii Balance Board®, Nintendo Co. Ltd., Kyoto, Japan). Four conditions were tested: bipedal standing with eyes open/closed, upright tandem standing with eyes open/closed. For each condition, the sway pattern, normalized path length, root mean square, and peak-to-peak values in medio-lateral and anterior–posterior direction were analyzed.
MRI protocol
The standardized protocol included whole brain and brainstem fine slice (3 mm) diffusion-weighted images (DWI), whole brain fluid attenuated inversion recovery (FLAIR)- and T2-weighted images including brainstem fine slicing (3 mm), T2*-weighted images, 3D-T1-weighted sequences (FSPGR 1 mm isovoxel) and time-of-flight angiography. All images were evaluated for the presence of ischemic stroke or bleeding by two specialized neuro-radiologists.
Classification methods
We prospectively evaluated two established diagnostic index tests, the HINTS and ABCD2 clinical scores for stroke detection, to establish a baseline classification performance. We compared these baselines against the performance of various modern machine-learning techniques. The latter learn the mapping of 305 input features (from history taking, questionnaires, and instrumentation-based examinations) to the output class of stroke vs. peripheral AVS. The classification performance is quantified with three diagnostic test measures [20], namely the area-under-the-curve of a receiver-operating-characteristic (ROC-AUC), accuracy, and F1-score, defined as:
Here, TP/TN/FP/FN indicate the number of true-positive/true-negative/false-positive/false-negative detections, respectively, and N indicates the number of test samples overall. The established diagnostic index tests and each of the machine-learning techniques are described briefly in the following.
HINTS: The HINTS clinical scoring system aggregates a risk score for detection of vestibular stroke, as proposed in [9]. HINTS constitutes a 3-step examination, based on Head Impulse, gaze-evoked Nystagmus, and Test of Skew. HINTS indicates a central pattern, if horizontal head impulse test is normal, and/or a direction-changing nystagmus in eccentric gaze, and/or a skew deviation is detected. Consequently, in our data set we give 1 point per central HINTS item and define a HINTS score cutoff value of ≥ 1 as indicative for vestibular stroke. From this binary value for stroke diagnosis, we compute the detection accuracy and F1-score. Additionally, we perform a receiver-operator-characteristic (ROC) analysis, varying the HINTS cutoff over our data set, to obtain an area-under-the-curve (AUC) score.
ABCD2: ABCD2 is an aggregative scoring system for clinical detection of stroke as proposed in [21] and validated in [22]. ABCD2 is based on the following features: age ≥ 60 years (1 point); blood pressure ≥ 140/90 mm Hg (1 point); clinical features: unilateral weakness (2 points), speech impairment without weakness (1 point); duration ≥ 60 min (2 points) or 10–59 min (1 point); and diabetes (1 point). For stroke detection in our study, we consider ABCD2 scores at a cutoff value of ≥ 3. We apply this cutoff to our dataset prospectively, and obtain the accuracy and F1-score, as well as a ROC-AUC score.
Logistic Regression (LR): In descriptive statistics, LR is used to report the goodness-of-fit of a linear set of equations, mapping a set of input features (i.e., observations) to a binary descriptor variable (e.g., stroke indicator variable). In this work, we use LR in a prospective/predictive manner. We regularize LR with a combined L1 and L2 loss, which allows learning of a Lasso-like sparse model, while still maintaining the regularization properties of a ridge classifier [23, 24]. The balancing ratio between the L1 and L2 losses is optimized during learning as a hyper-parameter. After fitting the LR parameters to samples in a training set, we apply the fitted model to samples in a holdout test set, to obtain a logistical posterior probability of stroke. We binarize the soft decision output of LR at a posterior probability , from which accuracy and F1-score are calculated. The AUC value is obtained by computing an ROC analysis on the probabilistic predictions for all samples.
Random Forest (RF): RF bundles an ensemble of decision tree (DT) models to compensate for tree overfitting [25] by vote aggregation [26]. In this work, we tune the number of DTs within the range of 5 to 50 trees towards optimal prediction performance. Due to the vote aggregation from the ensemble, an RF yields a probabilistic posterior. Accuracy, F1-score, and ROC-AUC are calculated on this posterior.
Artificial neural network (ANN): Computer-aided diagnosis has advanced due to the application of machine-learning techniques [27]. In particular, our own previous work [28–30], as well as numerous works in related literature [31] have demonstrated the effectiveness and modeling flexibility of ANNs for computer-aided diagnosis in medicine. Here, we apply a multilayer perceptron (MLP) with 305 input neurons, two hidden layers (128 and 64 neurons each), and two softmax-activated output neurons for classification. Due to the non-linear activation at the output layer, our ANN also yields a probabilistic posterior, allowing the calculation of accuracy, F1-score and ROC-AUC.
Geometric matrix completion (GMC): Geometric deep learning [32] is a novel field of deep learning, and has been introduced for computer-aided diagnosis in medicine only recently [33]. In previous work, we have shown that it is advantageous to construct multiple population graphs from meta-features of patients [34, 35]. We further proposed GMC [36] (denoted in the following as SingleGMC) to alleviate the common problem of missing values in medical data sets [37]. Recently, we have combined these ideas into multi-graph matrix completion (MultiGMC) [38]. Here, we apply both the original SingleGMC approach [36] and MultiGMC to our data set. In SingleGMC, we used a single graph and constructed it using age and ABCD2 scores. Graph connections are calculated based on similarity measures using age (age difference ± 5 years) and ABCD2 scores (± 1 score). For SingleGMC, the graph connectivity is the sum of these similarity measures. In MultiGMC, instead of taking the sum, we use them as two separate graphs. We learn separate patient representations within these two graphs (a single spectral convolutional layer per graph) and aggregate them via self-attention, before computing the classification posterior [38]. The calculation of accuracy, F1-score, and ROC-AUC is performed as for LR/RF/ANN.
The models LR, RF, and ANN were based on implementations in the scikit-learn machine-learning library [39], while GMC [36] and MultiGMC [38] are custom implementations, based on PyTorch [40].
Statistical analysis
Compared to HINTS and ABCD2, which are evaluated prospectively on the entire data set, the training of machine-learning models on the entire data set would result in overfitting and an overly optimistic performance estimate. Instead, we split the data into a training set and a test set, to obtain a prospective classification performance for our investigated models. All machine-learning based classification results were thus obtained following a rigorous ten-fold cross-validation scheme [41], with stratified label sampling to account for class imbalance, and a data split ratio of 90% training vs. 10% testing data. To perform hyper-parameter tuning for all methods, we monitored the tuned model performances on a withheld validation set (10% of the training set). We compared the best-performing model to the other four models, in terms of classification accuracy by pair-wise, two-tailed, non-parametric hypothesis tests (Wilcoxon signed-rank test) at a level of significance p < 0.05.
Furthermore, to make the results of the machine-learning classifier more explainable, we used the RF classifier to compute, which features contribute the most towards the detection of stroke. Such analysis constitutes a fundamental technique in the domain of machine-learning interpretability [42]. Feature importance was calculated according to the Mean Decrease in Impurity (MDI) measure [43], as implemented in scikit-learn [39]. We ranked the discriminative power of features by sorting the MDI coefficients, and reported the top 10 most important features utilized by the RF during classification. For these features, univariate analysis of quantitative values was performed for patients with vestibular stroke and vestibular neuritis (% for categorical variables, mean ± SD for continuous variables). The parameters were compared between groups using either the Chi-square test or Mann–Whitney U-test applying a significance level of p < 0.05.
Results
Prospective evaluation of HINTS and ABCD2 diagnostic performance
In a prospective analysis, we validated the classification scores of HINTS and ABCD2 for detection of all vestibular strokes (AVS and non-AVS presentation) against peripheral AVS. In our data set, HINTS was able to detect all strokes with an accuracy of 72.7%, at a ROC-AUC of 0.71. In comparison, ABCD2 detected stroke with a lower accuracy of 45.4%, at a ROC-AUC of 0.58. We indicate these univariate baseline methods as dashed horizontal lines in Fig. 1, to which we compare our machine-learning based models. HINTS had a diagnostic accuracy of 82.8%, at a ROC-AUC of 0.86 for stroke with AVS, and a diagnostic accuracy of 66.7%, at a ROC-AUC of 0.54 for stroke without AVS. ABCD2 performed with an accuracy of 37.7 (ROC-AUC of 0.59) for stroke with AVS, and 38.6% (ROC-AUC of 0.62) for stroke without AVS.
Machine-learning models for vestibular stroke detection
The median accuracy of all machine-learning methods ranged between 52% (LR) and 82% (MultiGMC). Two models, the linear LR and the non-linear ANN, achieved lower classification accuracy than the univariate measures HINTS and ABCD2 for all vestibular strokes vs. peripheral AVS, while RF/SingleGMC/MultiGMC were able to achieve better accuracy (Fig. 1a). Similar results were obtained for AUC and F1-score (Fig. 1b, c). Notably, two methods (RF and MultiGMC) were also able to achieve perfect classification accuracy, F1-score, and ROC-AUC for one of the five cross-validation folds, while LR and ANN, achieved a zero (0.0) F1-score for one of five folds. In general, MultiGMC yields comparably stable results over all five folds, with consistently high accuracy, F1-score, and ROC-AUC values. Comparing machine-learning classifiers statistically, MultiGMC classifies significantly better than LR (p < 0.01), ANN (p < 0.05), and SingleGMC (p < 0.05), but not significantly better than RF (p = 0.69).
Feature importance ranking
We used RF to rank features according to their discriminative performance. The top 10 selected features can be seen in Table 1. Features from device-based measurements such as VOG, SVV testing, and posturography, were considered as single parameters (e.g., vHIT-gain right, vHIT-gain left) or in an aggregated manner (vHIT pathological or normal based on a gain cutoff of 0.7 or presence of refixation saccades). No posturographic or SVV features were selected by the RF classifier as being among the top 10 important features. Instead, two aggregated VOG features (vHIT pathological, presence of horizontal SPN) and eight VOG-based single features were identified (e.g., vHIT gain, gaze-evoked nystagmus left, right).
Table 1.
Rank in RF | Feature | Feature type | Vestibular neuritis | Vestibular stroke | P value |
---|---|---|---|---|---|
1 | vHIT pathological (gain < 0.7/refixation saccades) | VOG (aggregated) | 100% | 12.5%* | < 0.0001 |
2 | vHIT gain (right) | VOG (single feature) | 0.6 ± 0.3** | 0.9 ± 0.3 | < 0.0001 |
3 | Fixation suppression of VOR gain (horizontal) |
VOG (single feature) |
0.03 ± 0.03 | 0.09 ± 0.06 | < 0.0001 |
4 | Smooth pursuit gain (downward direction) |
VOG (single feature) |
0.75 ± 0.17 | 0.67 ± 0.2 | 0.01 |
5 | SPN present without fixation (horizontal) |
VOG (aggregated) |
95.5%*** | 47.5% | < 0.0001 |
6 | SPV of SPN (0° position, vertical component) |
VOG (single feature) |
2.0 ± 2.5°/s | 1.0 ± 1.5°/s | 0.09 |
7 | SPV of GEN (15° right, horizontal component) |
VOG (single feature) |
1.2 ± 1.5°/s | 0.4 ± 0.6°/s | 0.004 |
8 | SPV of SPN (0° position, horizontal component) |
VOG (single feature) |
4.7 ± 4.0°/s | 1.0 ± 1.0°/s | < 0.0001 |
9 | SPV of GEN (15° left, horizontal component) |
VOG (single feature) |
1.6 ± 2.5°/s | 0.3 ± 0.4°/s | 0.002 |
10 | STD of SPN amplitude (0° position, horizontal) |
VOG (single feature) |
2.3 ± 1.4° | 1.8 ± 0.8° | 0.0005 |
Quantification of the respective features (as % or mean ± STD) in patients with vestibular neuritis or vestibular stroke and statistical intergroup comparison (Mann–Whitney U test for features 2–4 and 6–10, Chi-square test for features 1 and 5) (right side). GEN gaze-evoked nystagmus, SPN spontaneous nystagmus, SPV slow phase velocity, STD standard deviation, VOG videooculography, VOR vestibulo-ocular reflex, vHIT video head impulse test; *vHIT was pathological in vestibular stroke lesions affecting the vestibular nucleus or medial longitudinal fascicle; **Gain is depicted for the affected side in vestibular neuritis; ***In three patients without apparent SPN, symptoms of vestibular neuritis had already started ≥ 3 days before VOG recording
Quantitative univariate analysis of the 10 most important features revealed significant intergroup differences for all but one feature (i.e., rank 6, vertical component of SPN in 0° position, p = 0.09) (Fig. 1). The following features discriminated best between groups: (1) vHIT was pathologic in 100% of patients with vestibular neuritis (gain: 0.6 ± 0.3 at affected side), but only in 12.5% of patients with vestibular stroke (gain: 0.9 ± 0.3) (p < 0.0001). (2) SPN was found more frequently in vestibular neuritis (95.5%) than in vestibular stroke (47.5%), and was more intense (horizontal SPV in 0° position: 4.7 ± 4.0°/s vs. 1.0 ± 1.0°/s) (p < 0.0001). (3) Fixation suppression of the VOR was abnormal in vestibular stroke (gain: 0.09 ± 0.06), but intact in vestibular neuritis (gain: 0.03 ± 0.03). SPN was suppressed by fixation in 94% of patients with vestibular neuritis. Ranking of feature importance by RF reflected clinically important parameters with significant intergroup differences.
Discussion
Analysis of various approaches for the detection of patients with vestibular stroke (with the clinical presentation of AVS or non-AVS) vs. patients with peripheral AVS due to vestibular neuritis revealed the following findings: HINTS achieves better classification than ABCD2 and two of the tested machine-learning methods (LR, ANN), but is not as accurate as the more modern tested machine-learning methods (RF, Single-/MultiGMC) for differentiation of all vestibular strokes against peripheral AVS. In the following, we discuss the methodological and clinical implications of these findings.
Comparison of the different methodological approaches
In the current study, we compared two established clinical classification scores (HINTS, ABCD2) to a number of machine-learning techniques, both classical methods (LR, RF, ANN) and deep learning techniques based on population-modeling with graphs (SingleGMC, MultiGMC). In terms of median accuracy and area-under-the-curve (AUC), all machine-learning classifiers outperformed the detection rate of stroke as indicated by ABCD2. Compared to HINTS, however, several machine-learning classifiers performed similarly (LR, ANN, SingleGMC), while only RF and MultiGMC were able to reliably outperform HINTS. For vestibular stroke with AVS, the diagnostic accuracy of HINTS was comparable to MultiGMC. From a methodological perspective, our results provide a reliable estimate of a potential prospective classification performance for future validation studies, due to the usage of a rigorous cross-validation scheme and hyper-parameter optimization of all machine-learning models. More training data in prospective studies may improve results further, as data set size is usually a limiting factor in machine-learning studies [41]. The RF models yielded satisfactory results, while deep learning models, particularly MultiGMC, were able to improve results further. In general, the possibility to incorporate a semantic population model built from disease-relevant meta-features in form of a graph is attractive from a clinical point of view. The efficacy of this approach in everyday life clinical scenarios needs to be further validated in future studies.
Clinical implications
There is increasing discussion about the use of computer-aided diagnostic support systems in the context of complex clinical scenarios. The differentiation of central and peripheral etiologies of acute vertigo and dizziness poses such a challenge. Established diagnostic algorithms such as HINTS perform very well for AVS, which accounts for about half of acute presentations of vertigo or dizziness [9, 10]. Stroke detection remains particularly difficult, if patients have non-AVS presentations, transient or mild symptoms [3]. Therefore, in the current study we analyzed all vestibular stroke patients (AVS, non-AVS) against peripheral AVS. In our data set, ABCD2 had a low diagnostic performance to indicate vestibular stroke and HINTS outperformed ABCD2. Nevertheless, for all vestibular stroke patients (AVS, non-AVS), the diagnostic accuracy of HINTS was lower than previously reported for AVS only [9]. Modern machine-learning techniques (such as MultiGMC) had the highest diagnostic accuracy in separating vestibular stroke from peripheral AVS. Interestingly, ranking of feature importance by machine-learning algorithms (such as RF) closely resembled existing clinical experience. The top two features are derived from head impulse testing (vHIT pathologic, vHIT gain). In accordance, HIT has been previously considered the most important component of HINTS with a 18-fold stroke probability if normal in presence of SPN [44]. Two other features (ranks 7, 9) are concerned with gaze-evoked nystagmus, which is also part of HINTS. Skew deviation was not included in the 10 top features, which may be due to its low rate of manifestation (present in only about one quarter of vestibular stroke patients) [45]. Intensity of SPN was weighted prominently (ranks 5, 6, 8, 10). An additional feature with a high importance was a disturbed fixation suppression of the VOR (rank 3). This sign is regularly found in cerebellar lesions involving the uvula, pyramis, nodulus, and flocculus, which are common in patients with vestibular stroke [46, 47]. Notably, all the top-ranked features resulted from VOG examination, while SVV testing and posturography seemed to be less important. It is well-known that SVV deviation is found both in peripheral and central vestibular lesions, because it reflects a peripheral or central tone imbalance of graviceptive input originating from the vertical semicircular canals and otoliths [48]. The underrepresentation of postural parameters in our data set is in partial contrast to previous clinical studies, which have shown a high diagnostic relevance of the extent of falling tendency in AVS [49]. This discrepancy may be explained by the fact that the overall sway pattern cannot be derived from one or two features, but rather from a complex interplay of parameters [30].
Conclusions
This feasibility study shows the potential of modern machine-learning techniques to support diagnostic decisions in acute vestibular disorders. The current algorithm is tailored for the differentiation of vestibular neuritis vs. vestibular stroke only, and heavily depends on a quantitative and comprehensive assessment of vestibular and ocular motor functions by VOG, which limits its application under everyday life conditions in the ED. Therefore, future studies should focus on tailored VOG-protocols, include other qualitative factors (like triggers, acuity of onset, accompanying symptom features), and test the validity of machine-learning approaches in larger multicenter data sets for a wider range of differential diagnoses, such as Menière’s disease and vestibular migraine.
Acknowledgements
Open Access funding provided by Projekt DEAL. We thank Katie Göttlinger for copyediting of the manuscript.
Abbreviations
- ABCD2
Age, blood pressure, clinical features, duration, diabetes
- ANN
Artificial neural network
- AUC
Area-under-the-curve
- AVS
Acute vestibular syndrome
- CVRF
Cardiovascular risk factors
- DT
Decision tree
- DWI
Diffusion weighted images
- ED
Emergency department
- EMVERT
EMergency VERTigo
- FLAIR
Fluid attenuated inversion recovery
- GMC
Geometric matrix completion
- HINTS
Head impulse, gaze-evoked nystagmus, test of skew
- LR
Logistic regression
- ML
Machine-learning
- MLP
Multilayer perceptron
- MultiGMC
Multi-graph geometric matrix completion
- RF
Random forest
- ROC
Receiver operating characteristic
- SingleGMC
Single-graph geometric matrix completion
- SPN
Spontaneous nystagmus
- SPV
Slow phase velocity
- STD
Standard deviation
- SVV
Subjective visual vertical
- vHIT
Video-based head impulse test
- VOG
Videooculography
- VOR
Vestibulo-ocular reflex
Funding
The study was supported by the German Federal Ministry of Education and Research (BMBF) in connection with the foundation of the German Center for Vertigo and Balance Disorders (DSGZ) (Grant Number 01 EO 0901).
Compliance with ethical standards
Conflicts of interest
None of the authors has potential conflicts of interest to be disclosed.
Footnotes
Seyed-Ahmad Ahmadi and Gerome Vivar have contributed equally to this work.
References
- 1.Newman-Toker DE, Hsieh Y-H, Camargo CA, et al. Spectrum of dizziness visits to US emergency departments: cross-sectional analysis from a nationally representative sample. Mayo Clin Proc. 2008;83:765–775. doi: 10.4065/83.7.765. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Royl G, Ploner CJ, Leithner C. Dizziness in the emergency room: diagnoses and misdiagnoses. Eur Neurol. 2011;66:256–263. doi: 10.1159/000331046. [DOI] [PubMed] [Google Scholar]
- 3.Tarnutzer AA, Lee S-H, Robinson KA, et al. ED misdiagnosis of cerebrovascular events in the era of modern neuroimaging: a meta-analysis. Neurology. 2017;88:1468–1477. doi: 10.1212/WNL.0000000000003814. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Atzema CL, Grewal K, Lu H, et al. Outcomes among patients discharged from the emergency department with a diagnosis of peripheral vertigo: outcomes in patients discharged with peripheral vestibular disorders. Ann Neurol. 2016;79:32–41. doi: 10.1002/ana.24521. [DOI] [PubMed] [Google Scholar]
- 5.Saber Tehrani AS, Kattah JC, Mantokoudis G, et al. Small strokes causing severe vertigo: frequency of false-negative MRIs and nonlacunar mechanisms. Neurology. 2014;83:169–173. doi: 10.1212/WNL.0000000000000573. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Eagles D, Stiell IG, Clement CM, et al. International survey of emergency physicians’ priorities for clinical decision rules. Acad Emerg Med. 2008;15:177–182. doi: 10.1111/j.1553-2712.2008.00035.x. [DOI] [PubMed] [Google Scholar]
- 7.Zwergal A, Dieterich M. Vertigo and dizziness in the emergency room. Curr Opin Neurol. 2020;33:117–125. doi: 10.1097/WCO.0000000000000769. [DOI] [PubMed] [Google Scholar]
- 8.Choi K-D, Kim J-S. Vascular vertigo: updates. J Neurol. 2019;266:1835–1843. doi: 10.1007/s00415-018-9040-3. [DOI] [PubMed] [Google Scholar]
- 9.Kattah JC, Talkad AV, Wang DZ, et al. HINTS to diagnose stroke in the acute vestibular syndrome: three-step bedside oculomotor examination more sensitive than early MRI diffusion-weighted imaging. Stroke. 2009;40:3504–3510. doi: 10.1161/STROKEAHA.109.551234. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Newman-Toker DE, Tehrani ASS, Mantokoudis G, et al. Quantitative video-oculography to help diagnose stroke in acute vertigo and dizziness: toward an ECG for the eyes. Stroke. 2013;44:1158–1161. doi: 10.1161/STROKEAHA.111.000033. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Mantokoudis G, Saber Tehrani AS, Wozniak A, et al. VOR gain by head impulse video-oculography differentiates acute vestibular neuritis from stroke. Otol Neurotol. 2015;36:457–465. doi: 10.1097/MAO.0000000000000638. [DOI] [PubMed] [Google Scholar]
- 12.Navi BB, Kamel H, Shah MP, et al. Application of the ABCD 2 score to identify cerebrovascular causes of dizziness in the emergency department. Stroke. 2012;43:1484–1489. doi: 10.1161/STROKEAHA.111.646414. [DOI] [PubMed] [Google Scholar]
- 13.Newman-Toker DE, Kerber KA, Hsieh Y-H, et al. HINTS outperforms ABCD2 to screen for stroke in acute continuous vertigo and dizziness. Acad Emerg Med. 2013;20:986–996. doi: 10.1111/acem.12223. [DOI] [PubMed] [Google Scholar]
- 14.Choi J-H, Oh EH, Park M-G, et al. Early MRI-negative posterior circulation stroke presenting as acute dizziness. J Neurol. 2018;265:2993–3000. doi: 10.1007/s00415-018-9097-z. [DOI] [PubMed] [Google Scholar]
- 15.Möhwald K, Bardins S, Müller H-H, et al. Protocol for a prospective interventional trial to develop a diagnostic index test for stroke as a cause of vertigo, dizziness and imbalance in the emergency room (EMVERT study) BMJ Open. 2017;7:e019073. doi: 10.1136/bmjopen-2017-019073. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Herdman M, Gudex C, Lloyd A, et al. Development and preliminary testing of the new five-level version of EQ-5D (EQ-5D-5L) Qual Life Res. 2011;20:1727–1736. doi: 10.1007/s11136-011-9903-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Jacobson GP, Newman CW. The development of the dizziness handicap inventory. Arch Otolaryngol Head Neck Surg. 1990;116:424–427. doi: 10.1001/archotol.1990.01870040046011. [DOI] [PubMed] [Google Scholar]
- 18.Zwergal A, Rettinger N, Frenzel C, et al. A bucket of static vestibular function. Neurology. 2009;72:1689–1692. doi: 10.1212/WNL.0b013e3181a55ecf. [DOI] [PubMed] [Google Scholar]
- 19.Dieterich M, Brandt T. Perception of verticality and vestibular disorders of balance and falls. Front Neurol. 2019;10:172. doi: 10.3389/fneur.2019.00172. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Fawcett T. An introduction to ROC analysis. Pattern Recogn Lett. 2006;27:861–874. doi: 10.1016/j.patrec.2005.10.010. [DOI] [Google Scholar]
- 21.Johnston SC, Rothwell PM, Nguyen-Huynh MN, et al. Validation and refinement of scores to predict very early stroke risk after transient ischaemic attack. The Lancet. 2007;369:283–292. doi: 10.1016/S0140-6736(07)60150-0. [DOI] [PubMed] [Google Scholar]
- 22.Josephson SA, Sidney S, Pham TN, et al. Higher ABCD2 score predicts patients most likely to have true transient ischemic attack. Stroke. 2008;39:3096–3098. doi: 10.1161/STROKEAHA.108.514562. [DOI] [PubMed] [Google Scholar]
- 23.Friedman J, Hastie T, Tibshirani R (2010) Regularization paths for generalized linear models via coordinate descent. J Stat Soft. 10.18637/jss.v033.i01 [PMC free article] [PubMed]
- 24.Kim S-J, Koh K, Lustig M, et al. An interior-point method for large-scale -regularized least squares. IEEE J Sel Top Signal Process. 2007;1:606–617. doi: 10.1109/JSTSP.2007.910971. [DOI] [Google Scholar]
- 25.Hastie T, Tibshirani R, Friedman J. The elements of statistical learning. New York: Springer; 2009. [Google Scholar]
- 26.Criminisi A, Konukoglu E, Shotton J. Decision forests for classification, regression, density estimation. Microsoft Technical Report: Manifold Learning and Semi-Supervised Learning; 2011. [Google Scholar]
- 27.Yanase J, Triantaphyllou E. A systematic survey of computer-aided diagnosis in medicine: past and present developments. Expert Syst Appl. 2019;138:112821. doi: 10.1016/j.eswa.2019.112821. [DOI] [Google Scholar]
- 28.Krafczyk S, Tietze S, Swoboda W, et al. Artificial neural network: a new diagnostic posturographic tool for disorders of stance. Clin Neurophysiol. 2006;117:1692–1698. doi: 10.1016/j.clinph.2006.04.022. [DOI] [PubMed] [Google Scholar]
- 29.Pradhan C, Wuehr M, Akrami F, et al. Automated classification of neurological disorders of gait using spatio-temporal gait parameters. J Electromyograph Kinesiol. 2015;25:413–422. doi: 10.1016/j.jelekin.2015.01.004. [DOI] [PubMed] [Google Scholar]
- 30.Ahmadi S-A, Vivar G, Frei J, et al. Towards computerized diagnosis of neurological stance disorders: data mining and machine learning of posturography and sway. J Neurol. 2019;266:108–117. doi: 10.1007/s00415-019-09458-y. [DOI] [PubMed] [Google Scholar]
- 31.Lin D, Vasilakos AV, Tang Y, Yao Y. Neural networks for computer-aided diagnosis in medicine: a review. Neurocomputing. 2016;216:700–708. doi: 10.1016/j.neucom.2016.08.039. [DOI] [Google Scholar]
- 32.Bronstein MM, Bruna J, LeCun Y, et al. Geometric deep learning: going beyond Euclidean data. IEEE Signal Process Mag. 2017;34:18–42. doi: 10.1109/MSP.2017.2693418. [DOI] [Google Scholar]
- 33.Parisot S, Ktena SI, Ferrante E, et al. Disease prediction using graph convolutional networks: application to autism spectrum disorder and Alzheimer’s disease. Med Image Anal. 2018;48:117–130. doi: 10.1016/j.media.2018.06.001. [DOI] [PubMed] [Google Scholar]
- 34.Kazi A, Shekarforoush S, Arvind Krishna S, et al. et al. Graph convolution based attention model for personalized disease prediction. In: Shen D, Liu T, Peters TM, et al.et al., editors. Medical image computing and computer assisted intervention – MICCAI 2019. Cham: Springer; 2019. pp. 122–130. [Google Scholar]
- 35.Kazi A, Krishna SA, Shekarforoush S et al (2019) Self-attention equipped graph convolutions for disease prediction. In: 2019 IEEE 16th international symposium on biomedical imaging (ISBI 2019). IEEE, Venice, Italy, pp 1896–1899
- 36.Vivar G, Zwergal A, Navab N, Ahmadi S-A, et al. Multi-modal Disease Classification in Incomplete Datasets Using Geometric Matrix Completion. In: Stoyanov D, Taylor Z, Ferrante E, et al., editors. Graphs in Biomedical Image Analysis and Integrating Medical Imaging and Non-Imaging Modalities. Cham: Springer International Publishing; 2018. pp. 24–31. [Google Scholar]
- 37.Little RJ, D’Agostino R, Cohen ML, et al. The Prevention and Treatment of Missing Data in Clinical Trials. N Engl J Med. 2012;367:1355–1360. doi: 10.1056/NEJMsr1203730. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Vivar G, Kazi A, Zwergal A, et al Simultaneous imputation and disease classification in incomplete medical datasets using Multigraph Geometric Matrix Completion (MGMC). arXiv:2005.06935v1 [cs.LG] [DOI] [PubMed]
- 39.Pedregosa F, Varoquaux G, Gramfort A, et al. Scikit-learn: machine learning in python. Journal of Machine Learning Research. 2011;12:2825–2830. [Google Scholar]
- 40.Paszke A, Gross S, Massa F, et al (2019) PyTorch: an imperative style, high-performance deep learning library. In: Wallach H, Larochelle H, Beygelzimer A, et al (eds) Advances in neural information processing systems 32. Curran Associates, Inc., pp 8026–8037
- 41.Bishop CM (2006) Pattern recognition and machine learning (information science and statistics). Springer, New York
- 42.Molnar C (2019) Interpretable machine learning: a guide for making black box models explainable
- 43.Breiman L, Friedman JH, Olshen RA, Stone CJ (1984) Classification and regression trees, ed. 1. Chapman and Hall/CRC, New York
- 44.Tarnutzer AA, Berkowitz AL, Robinson KA, et al. Does my dizzy patient have a stroke? A systematic review of bedside diagnosis in acute vestibular syndrome. Can Med Assoc J. 2011;183:E571–E592. doi: 10.1503/cmaj.100174. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Brandt T, Dieterich M. The dizzy patient: don’t forget disorders of the central vestibular system. Nature Reviews Neurology. 2017;13:352–362. doi: 10.1038/nrneurol.2017.58. [DOI] [PubMed] [Google Scholar]
- 46.Baier B, Stoeter P, Dieterich M. Anatomical correlates of ocular motor deficits in cerebellar lesions. Brain. 2009;132:2114–2124. doi: 10.1093/brain/awp165. [DOI] [PubMed] [Google Scholar]
- 47.Zwergal A, Möhwald K, Salazar Lopez E et al (2020) A prospective analysis of lesion-symptom relationships in acute vestibular and ocular motor stroke. Front Neurol. (accepted) [DOI] [PMC free article] [PubMed]
- 48.Glasauer S, Dieterich M, Brandt T. Neuronal network-based mathematical modeling of perceived verticality in acute unilateral vestibular lesions: from nerve to thalamus and cortex. J Neurol. 2018;265:101–112. doi: 10.1007/s00415-018-8909-5. [DOI] [PubMed] [Google Scholar]
- 49.Carmona S, Martínez C, Zalazar G, et al. The diagnostic accuracy of truncal ataxia and HINTS as cardinal signs for acute vestibular syndrome. Front Neurol. 2016 doi: 10.3389/fneur.2016.00125. [DOI] [PMC free article] [PubMed] [Google Scholar]