Table 2.
Results of quantitative evaluation of features robustness of our SRHP algorithm against comparative methods at the slide-level in terms of AUC, Accuracy, Recall, Precision, Specificity, and F1 score. The results are averaged across 20 slides, which includes both G3 and G4 regions on the same slide. We also report standard deviations in performance metrics across slides (The best results are indicated in bold).
| Slide-level | AUC | Accuracy | Recall | Precision | Specificity | F1 Score |
|---|---|---|---|---|---|---|
| DLGg | 0.93 ± 0.13 | 89.35% ± 0.14 | 0.80 ± 0.25 | 0.82 ± 0.20 | 0.91 ± 0.13 | 0.78 ± 0.21 |
| SSAE | 0.62 ± 0.18 | 71.31% ± 0.12 | 0.42 ± 0.30 | 0.48 ± 0.36 | 0.73 ± 0.21 | 0.53 ± 0.25 |
| MATF | 0.94 ± 0.01 | 86.45% ± 0.02 | 0.84 ± 0.03 | 0.87 ± 0.05 | 0.89 ± 0.03 | 0.85 ± 0.03 |
| SRHP | 0.99 ± 0.01 | 98.75% ± 0.01 | 0.97 ± 0.03 | 0.93 ± 0.12 | 0.98 ± 0.02 | 0.95 ± 0.09 |