Recent Editorials in this journal stressed the classical paradigm in clinical epidemiology of insisting on test-retest evaluations for studies on diagnosis and prognosis [1], and specifically prediction models [2]. Indeed, independent validation of previous research findings is a general scientific principle.
Another recent debate was on the interpretation of the lack of external validation studies of published novel prediction models [3][4][5]. One issue is the role that validation should have at the time of model development. Many researchers may be tempted to try to report some proof for external validity, i.e. on discrimination and calibration, in independent samples with their publication that proposes a new prediction model. Major clinical journals currently seem to appreciate such reporting. Another issue is whether external validation should be performed by different authors than those involved in the development of the prediction model [3][6]. We would like to comment on these and related key issues in the scientific basis of prediction modeling.
The recent review confirms that model development studies are often relatively small for the complex challenges posed by specifying the form of a prediction model (which predictors to include) and the estimation of predictor effects (overfit with standard estimation methods) [3]. The median sample size was 445 subjects. The number of events is the limiting factor in this type of research and may be far too low for reliable modeling [4]. In such small samples, internal validation is essential, and apparent performance estimates are severely optimistic (Figure 1). Bootstrapping is the preferred approach for validation of prediction models [7] [8] [9]. A bootstrap procedure should include all modeling steps for an honest assessment of model performance [10]. Specifically any model selection steps, such as variable selection, need to be repeated per bootstrap sample if used.
We recently confirmed that a split sample approach with 50% held out leads to models with a suboptimal performance, i.e. models with unstable and on average the same performance as obtained with half the sample size [11]. We hence strongly advise against random split sample approaches in small development samples. Split sample approaches can be used in very large samples, but again we advise against this practice, since overfitting is no issue if sample size is so large that a split sample procedure can be performed. Split sample approaches only work when not needed.
More relevant are attempts to obtain impressions of external validity: do model predictions hold true in different settings, e.g. in subjects from other centers, or subjects seen more recently? Here a non-random split can often be made in the development sample e.g. by year of diagnosis. For example we might validate a model on the most recent 1/3 of the sample held out from model development. Since the split is in time, this would qualify as a temporal external validation [6]. The disadvantages of a random split sample approach unfortunately equally hold here: a poorer model is developed (since on smaller sample size than the full development sample), and the validation findings are unstable (since based on a small sample size) [9].
We make 2 propositions for validation at the time of prediction model development (Figure 2). First, we recommend an “internal- external” validation procedure. In the context of individual patient data meta-analysis (IPD-MA), internal-external cross-validation has been used to show external validity of a prediction model [12] [13]. In a MA context, the natural unit for splitting is by study. Every study is left out once, for validation of a model based on the remaining studies. The final model is based on the pooled data set, which we label an ‘internally-externally validated model’. In a more general sense, splits by hospital in a multicenter study or by calendar time are attractive options when developing prediction models, with the final model developed on all available data [14].
Second, we may consider more direct tests for heterogeneity in predictor effects by place or time. In IPD-MA with many studies, random effects models can be used to quantify the heterogeneity in predictor effects over studies [15]. The amount of heterogeneity is essential in claims on generalizability of predictions from the proposed model on the pooled data. In the same spirit we may assess heterogeneity over centers in a multicenter study where a prediction model is developed. When few studies are available, testing interaction terms such as “predictor * study” provide valuable insights, and more direct than a global impression of external validity. For temporal validity, a more direct test is provided by the interaction of “predictor * calender time”. In addition to predictor effects, such tests can also assess heterogeneity in baseline risk (main effect of “study”, after adjusting for risk), and in overall risk associations as summarized in the linear predictor (e.g. the “prognostic index * study” interaction). So again, using the maximal sample size in modeling is preferred.
Finally, fully independent external validation with data not available at the time of prediction model development can be important (Figure 2). Such validation should be a test of generalizability. If the external data set is very similar to the development data set, the assessment is for reproducibility rather than for transportability [6]. The similarity of validation to development sets hence is essential for interpretation of an external validation study; either by comparing descriptive data (“Table 1”) or a simple statistical model to predict membership of the development or validation data set [16].
In sum, we recommend that internal validation should always be attempted for any proposed prediction model, with bootstrapping being preferred (Figure 1). Many failed external validations could have been foreseen by rigorous internal validation, saving time and resources [4]. With respect to external validation at the time of model development, we recommend internal- external validation procedures and direct tests for heterogeneity of predictor effects rather than keeping parts of the data out (Figure 2). Such assessments at model development may temper overoptimistic expectations of prediction model performance in independent data. For external validation after the developed model was published, the issue may not so much be that different authors perform the analysis [5]; indeed, no difference was noted between validations performed by overlapping or different authors [3], although selective reporting and publication bias cannot be ruled out. What matters more is the similarity between settings [6][16], and the quality of the validation analysis [7] [8] [17], which may be positively influenced by reporting guidelines such as the recently proposed TRIPOD guidelines for prediction models [18] [19].
Acknowledgments
The work by Ewout Steyerberg was supported by a U award (AA022802, value of personalized risk information).
Footnotes
Publisher's Disclaimer: This is a PDF file of an unedited manuscript that has been accepted for publication. As a service to our customers we are providing this early version of the manuscript. The manuscript will undergo copyediting, typesetting, and review of the resulting proof before it is published in its final citable form. Please note that during the production process errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain.
There are no conflicts of interests for both authors.
Contributor Information
Ewout W. Steyerberg, Dept of Public Health, Erasmus MC, Rotterdam, The Netherlands
Frank E. Harrell, Jr, Department of Biostatistics, School of Medicine, Vanderbilt University, Nashville, TN, USA.
References
- 1.Tugwell P, Knottnerus JA. Transferability/generalizability deserves more attention in ‘retest’ studies in Diagnosis and Prognosis. J Clin Epidemiol. 2015 Mar;68(3):235–6. doi: 10.1016/j.jclinepi.2015.01.007. [DOI] [PubMed] [Google Scholar]
- 2.Tugwell P, Knottnerus JA. Clinical prediction models are not being validated. J Clin Epidemiol. 2015 Jan;68(1):1–2. doi: 10.1016/j.jclinepi.2014.11.020. [DOI] [PubMed] [Google Scholar]
- 3.Siontis GC, Tzoulaki I, Castaldi PJ, Ioannidis JP. External validation of new risk prediction models is infrequent and reveals worse prognostic discrimination. J Clin Epidemiol. 2015 Jan;68(1):25–34. doi: 10.1016/j.jclinepi.2014.09.007. [DOI] [PubMed] [Google Scholar]
- 4.Martens FK, Kers JG, Janssens AC. External validation is only needed when prediction models are worth it. J Clin Epidemiol. 2015 Feb 3; doi: 10.1016/j.jclinepi.2015.01.022. [DOI] [PubMed] [Google Scholar]
- 5.Siontis GC, Ioannidis JP. Response to letter: more rigorous, not less, external validation is needed. J Clin Epidemiol. 2015 Jan 31; doi: 10.1016/j.jclinepi.2015.01.021. [DOI] [PubMed] [Google Scholar]
- 6.Justice AC, Covinsky KE, Berlin JA. Assessing the generalizability of prognostic information. Ann Intern Med. 1999 Mar 16;130(6):515–24. doi: 10.7326/0003-4819-130-6-199903160-00016. [DOI] [PubMed] [Google Scholar]
- 7.Harrell FE. Regression Modeling Strategies: With Applications to Linear Models, Logistic Regression, and Survival Analysis. Springer; NY: 2001. [Google Scholar]
- 8.Steyerberg EW. Clinical Prediction Models: A Practical Approach to Development, Validation, and Updating. Springer; NY: 2009. [Google Scholar]
- 9.Steyerberg EW, Harrell FE, Jr, Borsboom GJ, Eijkemans MJ, Vergouwe Y, Habbema JD. Internal validation of predictive models: efficiency of some procedures for logistic regression analysis. J Clin Epidemiol. 2001 Aug;54(8):774–81. doi: 10.1016/s0895-4356(01)00341-9. [DOI] [PubMed] [Google Scholar]
- 10.Steyerberg EW, Bleeker SE, Moll HA, Grobbee DE, Moons KG. Internal and external validation of predictive models: a simulation study of bias and precision in small samples. J Clin Epidemiol. 2003 May;56(5):441–7. doi: 10.1016/s0895-4356(03)00047-7. [DOI] [PubMed] [Google Scholar]
- 11.Austin PC, Steyerberg EW. Events per variable (EPV) and the relative performance of different strategies for estimating the out-of-sample validity of logistic regression models. Stat Methods Med Res. 2014 Nov 19; doi: 10.1177/0962280214558972. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Royston P, Parmar MK, Sylvester R. Construction and validation of a prognostic model across several studies, with an application in superficial bladder cancer. Stat Med. 2004 Mar 30;23(6):907–26. doi: 10.1002/sim.1691. [DOI] [PubMed] [Google Scholar]
- 13.Steyerberg EW, Mushkudiani N, Perel P, Butcher I, Lu J, McHugh GS, Murray GD, Marmarou A, Roberts I, Habbema JD, Maas AI. Predicting outcome after traumatic brain injury: development and international validation of prognostic scores based on admission characteristics. PLoS Med. 2008 Aug 5;5(8):e165. doi: 10.1371/journal.pmed.0050165. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Balmaña J, Stockwell DH, Steyerberg EW, Stoffel EM, Deffenbaugh AM, Reid JE, Ward B, Scholl T, Hendrickson B, Tazelaar J, Burbidge LA, Syngal S. Prediction of MLH1 and MSH2 mutations in Lynch syndrome. JAMA. 2006 Sep 27;296(12):1469–78. doi: 10.1001/jama.296.12.1469. [DOI] [PubMed] [Google Scholar]
- 15.Legrand C, Duchateau L, Janssen P, Ducrocq V, Sylvester R. Validation of prognostic indices using the frailty model. Lifetime Data Anal. 2009 Mar;15(1):59–78. doi: 10.1007/s10985-008-9092-2. [DOI] [PubMed] [Google Scholar]
- 16.Debray TP, Vergouwe Y, Koffijberg H, Nieboer D, Steyerberg EW, Moons KG. A new framework to enhance the interpretation of external validation studies of clinical prediction models. J Clin Epidemiol. 2015 Mar;68(3):279–89. doi: 10.1016/j.jclinepi.2014.06.018. [DOI] [PubMed] [Google Scholar]
- 17.Steyerberg EW, Vergouwe Y. Towards better clinical prediction models: seven steps for development and an ABCD for validation. Eur Heart J. 2014 Aug 1;35(29):1925–31. doi: 10.1093/eurheartj/ehu207. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Collins GS, Reitsma JB, Altman DG, Moons KG. Transparent reporting of a multivariable prediction model for Individual Prognosis or Diagnosis (TRIPOD): the TRIPOD statement. J Clin Epidemiol. 2015 Feb;68(2):134–43. doi: 10.1016/j.jclinepi.2014.11.010. [DOI] [PubMed] [Google Scholar]
- 19.Moons KG, Altman DG, Reitsma JB, Ioannidis JP, Macaskill P, Steyerberg EW, Vickers AJ, Ransohoff DF, Collins GS. Transparent Reporting of a multivariable prediction model for Individual Prognosis or Diagnosis (TRIPOD): explanation and elaboration. Ann Intern Med. 2015 Jan 6;162(1):W1–73. doi: 10.7326/M14-0698. [DOI] [PubMed] [Google Scholar]