Table 3.
Technique | Rationale/use-case | Examples of implementation/approaches |
---|---|---|
Unsupervised learning | ||
Robustness | Unsupervised learning solutions robust to noise are reasonably (more) generalizable and less overfit to the trained-on data |
▪ Bootstrapping ▪ Sensitivity analysis |
Clinical interpretability | In unsupervised learning, algorithms may derive clusters/trajectories that differ by statistical measures, but are so similar that they are not clinically distinguishable. Assessing clinically the characteristics of subgroups is highly relevant. In supervised learning, this objective is less obvious, but clinically sensible predictor variables driving predictions most strongly could indicate a sensible/practically useful solution |
▪ Tables and other textual representation ▪ Figures, e.g., heatmaps, line plots etc |
Homogeneity within clusters and dissimilarity across clusters | The classical definition of well-defined clusters/trajectories, i.e,. that the subjects within the derived subgroups are similar to each other and dissimilar to subjects in the other subgroups). From a clinical perspective, this may also be done with e.g., heatmaps |
▪ Silhouette score ▪ Calinski-Harabasz score ▪ Davies-Bouldin score |
Model fit / complexity | Statistical measures of how well a model fits the observed data (with varying penalty for complexity of the solution) provides an easily interpretable metric for selecting the optimal model in a data-driven manner |
▪ Akaike information criterion (AIC; commonly suggests more complex solutions and larger number of subgroups) ▪ Bayesian information criterion (BIC; in contrast to AIC emphasizes simpler solutions [149]) |
Supervised learning | ||
Performance metrics | Various measures are available to assess how accurately the model predicts based on new input data, varying widely depending on the learning task and data at hand. Ideally, multiple metrics need to be evaluated to assess model performance, some being more suitable in specific types of prediction models [148, 150, 151] |
▪ R2, adjusted R2, MAE, RMSE, MAPE etc. (regression) ▪ MCC, F1 score, Cohen’s kappa, Brier score, AUROC, AUPRC, log-loss, accuracy, recall, specificity, NPV/PPV etc. (classification) |
Validation curve(s) | Diagnostic tool to assess the performance of a model in relation to hyperparameter settings (for example number of trees in a random forest model). More specifically, this can be used to e.g., evaluate hyperparameters that influences the model’s tendency to overfit/underfit | ▪ Plot of performance metric on y-axis and hyperparameter on x-axis |
The table describe common techniques used in model evaluation. The list is not meant to be comprehensive or universally applicable. It is recommended, particularly for unique applications, to evaluate previous similar implementations or relevant technical literature.