Skip to main content
The BMJ logoLink to The BMJ
. 2023 Feb 7;380:e071018. doi: 10.1136/bmj-2022-071018

Transparent reporting of multivariable prediction models developed or validated using clustered data: TRIPOD-Cluster checklist

Thomas P A Debray 1 ,2,, Gary S Collins 3 ,4, Richard D Riley 5, Kym I E Snell 5, Ben Van Calster 6, Johannes B Reitsma 1 ,2, Karel G M Moons 1 ,2
PMCID: PMC9903175  PMID: 36750242

Abstract

The increasing availability of large combined datasets (or big data), such as those from electronic health records and from individual participant data meta-analyses, provides new opportunities and challenges for researchers developing and validating (including updating) prediction models. These datasets typically include individuals from multiple clusters (such as multiple centres, geographical locations, or different studies). Accounting for clustering is important to avoid misleading conclusions and enables researchers to explore heterogeneity in prediction model performance across multiple centres, regions, or countries, to better tailor or match them to these different clusters, and thus to develop prediction models that are more generalisable. However, this requires prediction model researchers to adopt more specific design, analysis, and reporting methods than standard prediction model studies that do not have any inherent substantial clustering. Therefore, prediction model studies based on clustered data need to be reported differently so that readers can appraise the study methods and findings, further increasing the use and implementation of such prediction models developed or validated from clustered datasets.


The TRIPOD (transparent reporting of a multivariable prediction model for individual prognosis or diagnosis) statement provides guidance for the reporting of studies developing, validating, or updating a multivariable prediction model. However, the TRIPOD statement does not fully cover the specific methodological and analytical complexity in prediction model studies that are based on (and thus should account for) clustered data. We therefore developed TRIPOD-Cluster, a new reporting checklist for prediction model studies that are based on clustered datasets. Ten new items are introduced and 19 of the original TRIPOD items have been updated to better reflect issues for clustering. The rationale for each item is discussed, along with examples of good reporting and why transparent reporting is important. TRIPOD-Cluster is best used in conjunction with the TRIPOD-Cluster explanation and elaboration document (www.tripod-statement.org).

Prediction models combine a number of characteristics (predictors or covariates) to produce an individual’s probability or risk of currently having (diagnosis) or developing in a certain time period (prognosis) a specific condition or outcome.1 2 3 4 5 They are (still) typically derived using multivariable regression techniques,2 6 such as logistic regression (for diagnostic and short term binary prognostic outcomes), Cox proportional hazards regression (for long term binary outcomes), linear regression (for continuous outcomes), or multinomial regression (for categorical outcomes). However, they can be and, increasingly, are developed using machine learning algorithms such as neural networks, support vector machines, and random forests.7

Prediction models are widely used to support medical practice and diagnostic, prognostic, therapeutic, and preventive decision making.3 4 Therefore, prediction models should be developed, validated, and updated or tailored to different settings using appropriate methods and data sources, such that their predictions are sufficiently accurate and supportive for further decision making in the targeted population and setting.1 2 8 9 To evaluate whether a prediction model is fit for purpose, to properly assess their quality and any risks of bias, full and transparent reporting of prediction model studies is essential.10

The TRIPOD statement was published in 2015 to provide guidance on the reporting of prediction model studies.11 12 The TRIPOD statement comprises a checklist of 22 minimum reporting items that should be addressed in all prediction model studies, with translations available in Chinese and Japanese. However, the TRIPOD statement does not cover important items that arise when prediction models are based on clustered datasets.13 14 These datasets arise when individual participant data are obtained or combined from multiple groups, centres, settings, regions, countries, or studies (referred to here as clusters). Observations within a cluster tend to be more alike than observations from other clusters, leading to cluster populations that are different but related.

Clustered datasets can, for instance, be obtained by combining individual participant data from multiple studies (clustering by study; table 1), by conducting multicentre studies (clustering by centre), or by retrieving individual participant data from registries or datasets with electronic healthcare records (clustering by healthcare centre or geographical region; table 2). If differences between clusters are left unresolved during prediction model development, they might not perform well when applied in other or new clusters, and therefore have limited generalisability.29 30 Similarly, when heterogeneity between clusters is ignored during prediction model validation, estimates of prediction model performance can be highly misleading.31 32 33 In clustered or combined datasets, different individuals in the same cluster have been subject to, for example, similar healthcare processes that have also been delivered by the same healthcare providers, and therefore could be more alike than individuals from different clusters. Sometimes, clusters might also differ in participant eligibility criteria or participation, follow-up length, predictor and outcome definitions, and in (the quality of) applied measurement methods. Hence correlation is likely to be present between observations from the same data cluster,29 which can lead to differences or heterogeneity between clusters regarding patient characteristics, baseline risk, predictor effects, and outcome occurrence.

Table 1.

Examples of prediction model studies that are based on individual participant data meta-analysis from various medical domains and settings

IPD-MA project Population No of included studies Total sample size Type of prediction model study
IPPIC Network Pregnant women 78 ~3.6 million External validation of previously published prognostic prediction models for developing pre-eclampsia15
Emerging Risk Factors Collaboration People from the general population without known cardiovascular disease 120 ~1.8 million Development and validation of prediction models for predicting 10 year risk of fatal and non-fatal cardiovascular disease16 17
ZIKV IPD Consortium Pregnant women and their infants exposed to the Zika virus 42 ~30 000 Development and validation of prognostic models to predict the risk of miscarriage, fetal loss, and microcephaly18
Diagnostic individual patient data meta-analysis of published primary studies on the diagnosis of deep vein thrombosis Outpatients suspected for deep vein thrombosis 13 ~10 000 External validation study to assess the accuracy of the Wells rule for excluding deep vein thrombosis across different subgroups of patients19
IMPACT database of traumatic brain injury Patients with severe and moderate brain injuries 11 ~9000 Development of prognostic models to predict the risk of six month mortality and unfavourable outcomes20 21

IMPACT=International Mission for Prognosis And Clinical Trial; IPD-MA=individual participant data meta-analysis; IPPIC=International Prediction of Pregnancy Complication; ZIKV IPD=Zika virus individual participant data.

Table 2.

Examples of prediction model studies that are based on electronic healthcare records from various medical domains and settings

EHR database Population No of included clusters Total sample size Type of prediction model study
QResearch database Patients visiting a general practitioner 1309 general practices ~10.5 million Development and validation of the prognostic QRISK3 model to estimate the future risk of cardiovascular disease22 23
National ICD Registry Patients undergoing ICD implantation in routine clinical practice 1428 hospitals ~170 000 Development of a prediction model to estimate the risk for in-hospital adverse events among patients undergoing ICD placement24 25
CALIBER Patients visiting a general practitioner >300 general practices ~100 000 Development and validation of prognostic models for all-cause mortality and non-fatal myocardial infarction or coronary death in patients with stable coronary artery disease26 27
National Vascular Database Patients undergoing abdominal aortic aneurysm repair 140 hospitals ~11 000 Development of a risk prediction model for elective abdominal aortic aneurysm repair28

CALIBER=Cardiovascular disease research using Linked Bespoke studies and Electronic Health Records; EHR=electronic healthcare records; ICD=implantable cardioverter-defibrillators.

The presence of clustering has important benefits for prediction model research because it enables us to explore heterogeneity in prediction model performance (eg, the model’s calibration and discrimination) across, for example, multiple subgroups, centres, regions, or countries. Identifying sources of heterogeneity helps to better tailor or match the models to these different clusters, and thus to develop prediction models that are more generalisable. However, this strategy requires prediction model researchers to adopt more specific methods for design, analysis, and reporting than standard prediction model studies that do not have any inherent substantial clustering. Therefore, prediction model studies based on clustered data need to be reported differently, so that readers can appraise the study methods and findings, further increasing the use and implementation of prediction models developed or validated from clustered datasets. These specifics are not explicitly mentioned in the original TRIPOD statement.13 30 31

Similar to how CONSORT Cluster is an extension of CONSORT,32 we developed a new standalone prediction model reporting guideline, TRIPOD-Cluster, that provides guidance for transparent reporting of studies that describe the development or validation (including updating) of a diagnostic or prognostic prediction model using clustered data. TRIPOD-Cluster focuses on clustered studies as present in individual participant data meta-analysis (IPD-MA) and electronic healthcare records datasets (table 1 and table 2), and not on studies with clustering defined by repeated measurements within the same individuals. TRIPOD-Cluster is best used in conjunction with the TRIPOD-Cluster explanation and elaboration document. To aid the editorial process and readability of prediction model studies based on clustered data, it is recommended that authors include a completed checklist in their submission. All information is available online (www.tripod-statement.org).

Summary points.

  • To evaluate whether a prediction model is fit for purpose, and to properly assess their quality and any risks of bias, full and transparent reporting of prediction model studies is essential

  • The TRIPOD-Cluster statement is a new reporting checklist for prediction model studies that are based on clustered datasets

  • Clustered datasets can be obtained by combining individual participant data from multiple studies, by conducting multicentre studies, or by retrieving individual participant data from registries or datasets with electronic healthcare records

  • Presence of clustering can lead to differences (or heterogeneity) between clusters regarding participant characteristics, baseline risk, predictor effects, and outcome occurrence

  • Performance of prediction models can vary across clusters, and thereby affect their generalisability

  • Additional reporting efforts are needed in clustered data to clarify the identification of eligible data sources, data preparation, risk-of-bias assessment, heterogeneity in prediction model parameters, and heterogeneity in prediction model performance estimates


Predictions with clustered data: challenges and opportunities

Overfitting is a common problem in prediction model studies that use a single and often small dataset from a specific setting (eg, a single hospital or practice). Prediction models that are overfitted or overly simple tend to have poor accuracy when applied to new individuals from other hospitals or practices. In addition, when validation studies are also based on single, small, and local datasets, estimates of a prediction model’s performance likely become inaccurate and too optimistic, and fail to show generalisability across multiple medical practices, settings, or countries.33 Many examples of prediction models do not perform adequately when applied to new patients.34 35 36

To overcome these problems, interest in prediction model studies using large clustered datasets has grown. Combining data from multiple clusters has various advantages:

  • Prediction models developed, validated, or updated from larger combined datasets are less prone to overfitting, and estimates of prediction model performance are less likely to be optimistic.

  • Researchers can better investigate more complex but more predictive associations, such as including non-linearity of predictor effects, predictor interactions, and time varying effects.

  • The presence of multiple large clusters allows development and direct validation of models across a wide range of populations and settings, and thus allows direct tests of a model’s generalisability.13 37

  • Researchers can better explore and quantify any heterogeneity in a model’s predictive performance across multiple subgroups, centres, regions, or countries to better tailor or match the models to these different clusters, which is desirable when deciding on their use for individualised decision support.38

All these benefits of prediction modelling using clustered datasets improve model performance across new clusters, and help develop, validate, and update prediction models that are more generalisable across multiple types of clusters. Prediction model studies using clustered data include important additional details that are not explicitly detailed in the original TRIPOD statement.11 12 These details include varying predictor and outcome definitions across clusters, varying (quality of) predictor and outcome measurements, varying inclusion criteria or participation of individuals, the presence and handling of systematically missing data across and within clusters (ie, a predictor not measured for any individuals in a cluster), and varying design related aspects such as blinding. Accordingly, when a clustered dataset is used, special efforts in the development and validation of the prediction model might be needed during the analysis to avoid bias.

In general, adopting statistical methods that specifically account for clustering is recommended, because the presence of cluster effects could lead to differences in estimated model parameters and in the model’s predictive accuracy across clusters, such that ignoring clustering leads to limited generalisability and applicability.39 40 41 42 For instance, random effects models might be needed during prediction model development to allow for heterogeneity across clusters in baseline risk and predictor effects, which yields prediction models with varying intercepts and predictor effects for different clusters.39 40 Random effects models might also be needed during model validation to assess heterogeneity in prediction model performance across clusters to gain a deeper understanding of where and in which clusters the model does not satisfy.37 Other strategies to account for clustering might consider adjustment for cluster level variables, or implement variable selection algorithms that reduce the presence of cluster heterogeneity in model predictions.41 Strategies to explore the presence of clustering and account for statistical heterogeneity across clusters should be adequately reported, and are discussed in the TRIPOD-Cluster explanation and elaboration document.

None of the aspects mentioned above is included in the original TRIPOD checklist. For this reason, the present TRIPOD-Cluster checklist provides guidance for the reporting of studies describing the development, validation, and updating of a multivariable diagnostic or prognostic prediction model using clustered data. We provide a flow chart to help authors decide whether to use TRIPOD-Cluster or the original TRIPOD checklist (fig 1). We further refer to the explanation and elaboration document for detailed reasoning and guidance why a new cluster specific item has been added to the original TRIPOD checklist, in which we also use the example studies listed in table 1 and table 2.

Fig 1.

Fig 1

Flow chart to determine whether the published study should follow the original TRIPOD statement or TRIPOD-Cluster checklist

Development of TRIPOD-Cluster

We followed the strategy proposed by the EQUATOR (enhancing the quality and transparency of health research) Network (www.equator-network.org) for developing health research reporting guidelines and formed an executive group (consisting of Doug Altman, GC, TD, KM, JR, and RR).42

The first step consisted of reviewing the literature and the EQUATOR Network database (www.equator-network.org) to identify existing reporting guidelines for large (clustered) data or IPD-MA studies in general, including any extensions of these guidelines. Reporting guidelines have been published for various research designs, and these contain many items that are also relevant to prediction model studies using clustered data. In particular, the PRISMA (preferred reporting items for systematic reviews and meta-analyses) checklist of individual participant data (PRISMA-IPD),43 STROBE (strengthening the reporting of obeservational studies in epidemiology) statement,44 and RECORD (reporting of studies conducted using observational routinely collected data) guideline45 contain several items that are also relevant for the transparent reporting of prediction model studies using datasets from electronic healthcare records, multicentre cohort studies, or IPD-MA studies.

The executive group subsequently convened 11 times in person over four years to discuss the development and structure of TRIPOD-Cluster, the rationale for including new items, topics requiring further consideration, and the publication strategy. The initial face-to-face meetings were also used to initiate a large Delphi survey and to test the preceding versions of this TRIPOD-Cluster checklist.

A first draft of the TRIPOD-Cluster checklist was then prepared by the executive group. This draft formed the basis of an electronic questionnaire distributed to 77 Delphi participants, involving statisticians, epidemiologists, physicians, and journal editors with experience in prediction model research. The Delphi survey was opened in January 2019 and remained open for 50 days. A total of 27 modifications or new items to the original TRIPOD checklist were proposed to the Delphi participants, who rated each as “agree,” “no opinion,” or “disagree (please comment).” Three items had major disagreements (>30% scored “disagree”), and another three items had moderate disagreements (15-30% scored “disagree”), most of which were related to the (re)phrasing of newly proposed or existing items (appendix). Subsequently, a revised revision of the TRIPOD-Cluster checklist was prepared, and a second Delphi round among the same 77 participants was implemented in March 2019. All the attendant suggestions and results were then incorporated into the TRIPOD-Cluster checklist, which was discussed among participants of the annual conference of the International Society for Clinical Biostatistics (Leuven, 2019) to further refine the checklist.

How to use the TRIPOD-Cluster checklist and the main changes from the original TRIPOD checklist

The TRIPOD-Cluster checklist is not intended to be prescriptive about how prediction model studies using clustered data should be conducted or interpreted, although we believe that the detailed guidance for each item in the TRIPOD statement will help researchers and readers for this purpose. Although TRIPOD-Cluster is to be used as a standalone reporting guideline, we also recommend using TRIPOD-Cluster alongside the original TRIPOD explanation and elaboration document,12 and when applicable, the PROBAST (prediction model risk-of-bias assessment tool) guidance and its accompanying explanation and elaboration on the evaluation of risk bias in prediction models.46 47 When prediction model studies do not account for clustering (eg, because cluster identifiers were not recorded or clusters were too small), we recommend that authors explicitly report this and adopt the original TRIPOD checklist.

Reporting of all relevant information might not always be feasible, for instance, owing to word count limits of specific journals or websites. In these situations, researchers might summarise the most relevant information in a table or figure, and provide additional details in the supplementary material.

A total of 19 original TRIPOD items and subitems were modified. Some subitems from the original TRIPOD checklist were merged in TRIPOD-Cluster for simplification. These merged subitems include the definition and assessment of outcomes (TRIPOD items 6a and 6b) and predictors (TRIPOD items 7a and 7b), characteristics of included data sources (TRIPOD items 4a, 4b, 5a, 5b, 13b, and 14a), and details on the model specification (TRIPOD items 15a and 15b).

Finally, 10 new subitems were included to look at identification of multiple eligible data sources or clusters (eg, in IPD-MA studies), data preparation, risk-of-bias assessment, heterogeneity in prediction model parameters, heterogeneity in prediction model performance estimates, and cluster specific and sensitivity analyses.

The resulting TRIPOD-Cluster guidance comprises 19 main items and is presented in this manuscript (table 3; separate version in the supplementary table). An explanation for each item and subitem is described in the accompanying explanation and elaboration document, where each subitem is illustrated with examples of good reporting from published electronic healthcare records (or multicentre) and IPD-MA studies, including those listed in table 1 and table 2.

Table 3.

TRIPOD-Cluster checklist of items to include when reporting a study developing or validating a multivariable prediction model using clustered data

# Description Page #
Title and abstract
1 Identify the study as developing and/or validating a multivariable prediction model, the target population, and the outcome to be predicted.  
2 Provide a summary of research objectives, setting, participants, data source, sample size, predictors, outcome, statistical analysis, results, and conclusions.*  
Introduction
3a Explain the medical context (including whether diagnostic or prognostic) and rationale for developing or validating the prediction model, including references to existing models, and the advantages of the study design.*  
3b Specify the objectives, including whether the study describes the development or validation of the model.*  
Methods
4a Describe eligibility criteria for participants and datasets.*  
4b Describe the origin of the data, and how the data were identified, requested, and collected.  
5 Explain how the sample size was arrived at.*  
6a Define the outcome that is predicted by the model, including how and when assessed.*  
6b Define all predictors used in developing or validating the model, including how and when measured.*  
7a Describe how the data were prepared for analysis, including any cleaning, harmonisation, linkage, and quality checks.  
7b Describe the method for assessing risk of bias and applicability in the individual clusters (eg, using PROBAST).  
7c For validation, identify any differences in definition and measurement from the development data (eg, setting, eligibility criteria, outcome, predictors).*  
7d Describe how missing data were handled.*  
8a Describe how predictors were handled in the analyses.  
8b Specify the type of model, all model-building procedures (eg, any predictor selection and penalisation), and method for validation.*  
8c Describe how any heterogeneity across clusters (eg, studies or settings) in model parameter values was handled.  
8d For validation, describe how the predictions were calculated.  
8e Specify all measures used to assess model performance (eg, calibration, discrimination, and decision curve analysis) and, if relevant, to compare multiple models.  
8f Describe how any heterogeneity across clusters (eg, studies or settings) in model performance was handled and quantified.  
8g Describe any model updating (eg, recalibration) arising from the validation, either overall or for particular populations or settings.*  
9 Describe any planned subgroup or sensitivity analysis, (eg, assessing performance according to sources of bias, participant characteristics, setting).  
Results
10a Describe the number of clusters and participants from data identified through to data analysed. A flow chart may be helpful.*  
10b Report the characteristics overall and where applicable for each data source or setting, including the key dates, predictors, treatments received, sample size, number of outcome events, follow-up time, and amount of missing data.*  
10c For validation, show a comparison with the development data of the distribution of important variables (demographics, predictors, and outcome).  
11 Report the results of the risk of bias assessment in the individual clusters.  
12a Report the results of any across-cluster heterogeneity assessments that led to subsequent actions during the model’s development (eg, inclusion or exclusion of particular predictors or clusters).  
12b Present the final prediction model (ie, all regression coefficients, and model intercept or baseline estimate of the outcome at a given time point) and explain how to use it for predictions in new individuals.*  
13a Report performance measures (with uncertainty intervals) for the prediction model, overall and for each cluster.  
13b Report results of any heterogeneity across clusters in model performance.  
14 Report the results from any model updating (including the updated model equation and subsequent performance), overall and for each cluster.*  
15 Report results from any subgroup or sensitivity analysis.  
Discussion
16a Give an overall interpretation of the main results, including heterogeneity across clusters in model performance, in the context of the objectives and previous studies.*  
16b For validation, discuss the results with reference to the model performance in the development data, and in any previous validations.  
16c Discuss the strengths of the study and any limitations (eg, missing or incomplete data, non-representativeness, data harmonisation problems).*  
17 Discuss the potential use of the model and implications for future research, with specific view to generalisability and applicability of the model across different settings or (sub)populations.*  
Other information
18 Provide information about the availability of supplementary resources (eg, study protocol, analysis code, datasets).*  
19 Give the source of funding and the role of the funders for the present study.  

A separate version of this checklist is available in the supplementary table.

PROBAST=prediction model risk-of-bias assessment tool.

*

Item text is an adaptation of one or more existing items from the original TRIPOD (transparent reporting of a multivariable prediction model for individual prognosis or diagnosis) checklist.

Concluding remarks

TRIPOD-Cluster provides comprehensive consensus based guidance for the reporting of studies describing the development, validation, or updating of a multivariable diagnostic or prognostic prediction model using clustered data, such as those using electronic healthcare records or IPD-MA datasets (table 1 and table 2). TRIPOD-Cluster is not intended for reporting of studies in which the clustering is determined by repeated measurements of the same individuals. The checklist provides explicit guidance on the minimum required information to report, to help readers and potential model users understand how the study was designed, conducted, analysed, and inferred. We do not propose a standardised structure of reporting; rather, all the information requested in the checklist should be reported somewhere in the report or article, including supplementary material. We encourage authors to complete the checklist indicating the page number where the relevant information is reported and to submit this checklist with the article. The TRIPOD-Cluster checklist can be downloaded online (www.tripod-statement.org). Announcements relating to TRIPOD-Cluster will be made via www.tripod-statement.org and by the TRIPOD Twitter account (@TRIPODstatement). The EQUATOR Network will help disseminate and promote TRIPOD-Cluster.

Acknowledgments

We thank the members of the Delphi panel for their valuable input: Jeffrey Blume (Vanderbilt University, Nashville, TN, USA), Laura Bonnett (University of Liverpool, UK), Patrick Bossuyt (Amsterdam University Medical Centre, the Netherlands), Nancy Cook (Brigham and Women's Hospital, Boston, MA, USA), Carol Coupland (University of Nottingham, UK), Anneke Damen (University Medical Centre Utrecht, Utrecht University, the Netherlands), Joie Ensor (Keele University, UK), Frank Harrell (Vanderbilt University, Nashville, TN, USA), Georg Heinze (Medical University of Vienna, Austria), Martijn Heymans (Amsterdam University Medical Centre, the Netherlands), Lotty Hooft (Cochrane Netherlands, University Medical Centre Utrecht, Utrecht University, the Netherlands), Teus H Kappen (University Medical Centre Utrecht, Utrecht University, the Netherlands), Michael Kattan (Cleveland Clinic, Cleveland, OH, USA), Yannick Le Manach (McMaster University, Hamilton, ON, Canada), Glen Martin (University of Manchester, UK), Cynthia Mulrow (UT Health San Antonio Department of Medicine, San Antonio, TX, USA), Tri-Long Nguyen (University of Copenhagen, Denmark), Romin Pajouheshnia (Utrecht University, the Netherlands), Linda Peelen (University Medical Centre Utrecht, Utrecht University, the Netherlands), Willi Sauerbrei (University of Freiburg, Germany), Christopher Schmid (Brown University, Providence, RI, USA), Rob Scholten (Cochrane Netherlands, University Medical Centre Utrecht, Utrecht University, the Netherlands), Ewoud Schuit (University Medical Centre Utrecht, Utrecht University, the Netherlands), Ewout Steyerberg (Leiden University Medical Centre and Erasmus Medical Centre, the Netherlands), Yemisi Takwoingi (University of Birmingham, UK), Laure Wynants (Maastricht University, the Netherlands, and KU Leuven, Belgium).

We dedicate TRIPOD-Cluster to our friend Doug Altman, who co-initiated and was involved in TRIPOD-Cluster until his passing away on 3 June 2018.

Web extra.

Extra material supplied by authors

Web appendix: Supplementary information

debt071018.ww.pdf (457.5KB, pdf)

Supplementary table: TRIPOD-Cluster checklist

debt071018.wt.pdf (82.3KB, pdf)

Contributors: TPAD is the guarantor. TPAD, GSC, RDR, Doug Altman, JBR, and KGMM conceived the TRIPOD-Cluster statement and developed an initial checklist. TPAD organised two Delphi rounds to obtain feedback on individual items of the TRIPOD-Cluster checklist. TPAD and KGMM prepared the first draft of the manuscript. All authors reviewed subsequent drafts and assisted in revising the manuscript. The corresponding author attests that all listed authors meet authorship criteria and that no others meeting the criteria have been omitted.

Funding: No specific funding was given for this work. TPAD and JR acknowledge financial support from the Netherlands Organisation for Health Research and Development (grants 91617050 and 91215058). TPAD also acknowledges financial support from the European Union’s Horizon 2020 Research and Innovation Programme under ReCoDID Grant Agreement no. 825746. GSC is supported by the National Institute for Health and Care Research (NIHR) Biomedical Research Centre, Oxford, UK and Cancer Research UK (grant C49297/A27294). KIES is funded by the NIHR School for Primary Care Research (NIHR SPCR). BVC is supported by Research Foundation–Flanders (FWO; grants G0B4716N and G097322N) and Internal Funds KU Leuven (grants C24/15/037 and C24M/20/064). The views expressed are those of the author(s) and not necessarily those of the NHS, the NIHR, the Department of Health and Social Care, or any other funding body. The funders had no role in considering the study design or in the collection, analysis, interpretation of data, writing of the report, or decision to submit the article for publication.

Competing interests: All authors have completed the ICMJE uniform disclosure form at www.icmje.org/disclosure-of-interest/ and declare: GSC, KIES, BVC, JBR, and KGMM have no competing interests to disclose. RDR has received personal payments from universities and other organisations for providing training courses, and receives royalties for sales from two handbooks. TPAD provides consulting services to life science industries, including pharmaceutical companies and contract research organisations.

Provenance and peer review: Not commissioned; externally peer reviewed.

References

  • 1. Moons KGM, Royston P, Vergouwe Y, Grobbee DE, Altman DG. Prognosis and prognostic research: what, why, and how? BMJ 2009;338:b375. 10.1136/bmj.b375  [DOI] [PubMed] [Google Scholar]
  • 2. Steyerberg EW, Moons KGM, van der Windt DA, et al. PROGRESS Group . Prognosis Research Strategy (PROGRESS) 3: prognostic model research. PLoS Med 2013;10:e1001381. 10.1371/journal.pmed.1001381  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3. Riley RD, van der Windt D, Croft P, Moons KGM, eds. Prognosis Research in Healthcare: concepts, methods, and impact. 1st ed. Oxford University Press, 2019. 10.1093/med/9780198796619.001.0001. [DOI] [Google Scholar]
  • 4. Steyerberg EW. Clinical Prediction Models: A Practical Approach to Development, Validation, and Updating. 2nd ed. Springer International Publishing; 2019. 10.1007/978-3-030-16399-0. [DOI] [Google Scholar]
  • 5. Harrell FE. Regression modeling strategies: with applications to linear models, logistic and ordinal regression, and survival analysis. 2nd ed. Springer; 2015. 10.1007/978-3-319-19425-7. [DOI] [Google Scholar]
  • 6. Bouwmeester W, Zuithoff NPA, Mallett S, et al. Reporting and methods in clinical prediction research: a systematic review. PLoS Med 2012;9:1-12. 10.1371/journal.pmed.1001221  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7. Christodoulou E, Ma J, Collins GS, Steyerberg EW, Verbakel JY, Van Calster B. A systematic review shows no performance benefit of machine learning over logistic regression for clinical prediction models. J Clin Epidemiol 2019;110:12-22. 10.1016/j.jclinepi.2019.02.004  [DOI] [PubMed] [Google Scholar]
  • 8. Altman DG, Vergouwe Y, Royston P, Moons KGM. Prognosis and prognostic research: validating a prognostic model. BMJ 2009;338:b605. 10.1136/bmj.b605  [DOI] [PubMed] [Google Scholar]
  • 9. Royston P, Moons KGM, Altman DG, Vergouwe Y. Prognosis and prognostic research: Developing a prognostic model. BMJ 2009;338:b604. 10.1136/bmj.b604  [DOI] [PubMed] [Google Scholar]
  • 10. Peat G, Riley RD, Croft P, et al. PROGRESS Group . Improving the transparency of prognosis research: the role of reporting, data sharing, registration, and protocols. PLoS Med 2014;11:e1001671. 10.1371/journal.pmed.1001671  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11. Collins GS, Reitsma JB, Altman DG, Moons KGM. Transparent Reporting of a multivariable prediction model for Individual Prognosis or Diagnosis (TRIPOD): the TRIPOD statement. Ann Intern Med 2015;162:55-63. 10.7326/M14-0697  [DOI] [PubMed] [Google Scholar]
  • 12. Moons KGM, Altman DG, Reitsma JB, et al. Transparent Reporting of a multivariable prediction model for Individual Prognosis or Diagnosis (TRIPOD): explanation and elaboration. Ann Intern Med 2015;162:W1-73. 10.7326/M14-0698  [DOI] [PubMed] [Google Scholar]
  • 13. Debray TPA, Riley RD, Rovers MM, Reitsma JB, Moons KG, Cochrane IPD Meta-analysis Methods group . Individual participant data (IPD) meta-analyses of diagnostic and prognostic modeling studies: guidance on their use. PLoS Med 2015;12:e1001886. 10.1371/journal.pmed.1001886  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14. Cook JA, Collins GS. The rise of big clinical databases. Br J Surg 2015;102:e93-101. 10.1002/bjs.9723  [DOI] [PubMed] [Google Scholar]
  • 15. Allotey J, Snell KIE, Chan C, et al. IPPIC Collaborative Network . External validation, update and development of prediction models for pre-eclampsia using an Individual Participant Data (IPD) meta-analysis: the International Prediction of Pregnancy Complication Network (IPPIC pre-eclampsia) protocol. Diagn Progn Res 2017;1:16. 10.1186/s41512-017-0016-z  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16. WHO CVD Risk Chart Working Group . World Health Organization cardiovascular disease risk charts: revised models to estimate risk in 21 global regions. Lancet Glob Health 2019;7:e1332-45. 10.1016/S2214-109X(19)30318-3  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17. Danesh J, Erqou S, Walker M, et al. Emerging Risk Factors Collaboration . The Emerging Risk Factors Collaboration: analysis of individual data on lipid, inflammatory and other markers in over 1.1 million participants in 104 prospective studies of cardiovascular diseases. Eur J Epidemiol 2007;22:839-69. 10.1007/s10654-007-9165-7  [DOI] [PubMed] [Google Scholar]
  • 18. Wilder-Smith A, Wei Y, Araújo TVB, et al. Zika Virus Individual Participant Data Consortium . Understanding the relation between Zika virus infection during pregnancy and adverse fetal, infant and child outcomes: a protocol for a systematic review and individual participant data meta-analysis of longitudinal studies of pregnant women and their infants and children. BMJ Open 2019;9:e026092. 10.1136/bmjopen-2018-026092  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19. Geersing GJ, Zuithoff NPA, Kearon C, et al. Exclusion of deep vein thrombosis using the Wells rule in clinically important subgroups: individual patient data meta-analysis. BMJ 2014;348:g1340. 10.1136/bmj.g1340  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20. Steyerberg EW, Mushkudiani N, Perel P, et al. Predicting outcome after traumatic brain injury: development and international validation of prognostic scores based on admission characteristics. PLoS Med 2008;5:e165 , discussion e165. 10.1371/journal.pmed.0050165  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21. Marmarou A, Lu J, Butcher I, et al. IMPACT database of traumatic brain injury: design and description. J Neurotrauma 2007;24:239-50. 10.1089/neu.2006.0036  [DOI] [PubMed] [Google Scholar]
  • 22. Hippisley-Cox J, Stables D, Pringle M. QRESEARCH: a new general practice database for research. Inform Prim Care 2004;12:49-50. [DOI] [PubMed] [Google Scholar]
  • 23. Hippisley-Cox J, Coupland C, Brindle P. Development and validation of QRISK3 risk prediction algorithms to estimate future risk of cardiovascular disease: prospective cohort study. BMJ 2017;357:j2099. 10.1136/bmj.j2099  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24. Hammill SC, Stevenson LW, Kadish AH, et al. Review of the registry’s first year, data collected, and future plans. Heart Rhythm 2007;4:1260-3. 10.1016/j.hrthm.2007.07.021  [DOI] [PubMed] [Google Scholar]
  • 25. Dodson JA, Reynolds MR, Bao H, et al. NCDR . Developing a risk model for in-hospital adverse events following implantable cardioverter-defibrillator implantation: a report from the NCDR (National Cardiovascular Data Registry). J Am Coll Cardiol 2014;63:788-96. 10.1016/j.jacc.2013.09.079  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26. Denaxas SC, George J, Herrett E, et al. Data resource profile: cardiovascular disease research using linked bespoke studies and electronic health records (CALIBER). Int J Epidemiol 2012;41:1625-38. 10.1093/ije/dys188  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27. Rapsomaniki E, Shah A, Perel P, et al. Prognostic models for stable coronary artery disease based on electronic health record cohort of 102 023 patients. Eur Heart J 2014;35:844-52. 10.1093/eurheartj/eht533  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28. Grant SW, Hickey GL, Grayson AD, Mitchell DC, McCollum CN. National risk prediction model for elective abdominal aortic aneurysm repair. Br J Surg 2013;100:645-53. 10.1002/bjs.9047  [DOI] [PubMed] [Google Scholar]
  • 29. Localio AR, Berlin JA, Ten Have TR, Kimmel SE. Adjustments for center in multicenter studies: an overview. Ann Intern Med 2001;135:112-23. 10.7326/0003-4819-135-2-200107170-00012  [DOI] [PubMed] [Google Scholar]
  • 30. Riley RD, Tierney JF, Stewart LA, eds. Individual participant data meta-analysis: a handbook for healthcare research. Wiley; 2021. 10.1002/9781119333784. [DOI] [Google Scholar]
  • 31. Debray TPA, de Jong VMT, Moons KGM, Riley RD. Evidence synthesis in prognosis research. Diagn Progn Res 2019;3:13. 10.1186/s41512-019-0059-4  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32. Campbell MK, Elbourne DR, Altman DG, CONSORT group . CONSORT statement: extension to cluster randomised trials. BMJ 2004;328:702-8. 10.1136/bmj.328.7441.702  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33. Collins GS, Ogundimu EO, Altman DG. Sample size considerations for the external validation of a multivariable prognostic model: a resampling study. Stat Med 2016;35:214-26. 10.1002/sim.6787  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34. Allotey J, Snell KI, Smuk M, et al. Validation and development of models using clinical, biochemical and ultrasound markers for predicting pre-eclampsia: an individual participant data meta-analysis. Health Technol Assess 2020;24:1-252. 10.3310/hta24720  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35. Meads C, Ahmed I, Riley RD. A systematic review of breast cancer incidence risk prediction models with meta-analysis of their performance. Breast Cancer Res Treat 2012;132:365-77. 10.1007/s10549-011-1818-2  [DOI] [PubMed] [Google Scholar]
  • 36. Damen JA, Pajouheshnia R, Heus P, et al. Performance of the Framingham risk models and pooled cohort equations for predicting 10-year risk of cardiovascular disease: a systematic review and meta-analysis. BMC Med 2019;17:109. 10.1186/s12916-019-1340-7  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37. Debray TPA, Vergouwe Y, Koffijberg H, Nieboer D, Steyerberg EW, Moons KGM. A new framework to enhance the interpretation of external validation studies of clinical prediction models. J Clin Epidemiol 2015;68:279-89. 10.1016/j.jclinepi.2014.06.018  [DOI] [PubMed] [Google Scholar]
  • 38. Van Calster B, Nieboer D, Vergouwe Y, De Cock B, Pencina MJ, Steyerberg EW. A calibration hierarchy for risk models was defined: from utopia to empirical data. J Clin Epidemiol 2016;74:167-76. 10.1016/j.jclinepi.2015.12.005  [DOI] [PubMed] [Google Scholar]
  • 39. Debray TPA, Moons KGM, Ahmed I, Koffijberg H, Riley RD. A framework for developing, implementing, and evaluating clinical prediction models in an individual participant data meta-analysis. Stat Med 2013;32:3158-80. 10.1002/sim.5732  [DOI] [PubMed] [Google Scholar]
  • 40. Wynants L, Vergouwe Y, Van Huffel S, Timmerman D, Van Calster B. Does ignoring clustering in multicenter data influence the performance of prediction models? A simulation study. Stat Methods Med Res 2018;27:1723-36. 10.1177/0962280216668555  [DOI] [PubMed] [Google Scholar]
  • 41. de Jong VMT, Moons KGM, Eijkemans MJC, Riley RD, Debray TPA. Developing more generalizable prediction models from pooled studies and large clustered data sets. Stat Med 2021;40:3533-59. 10.1002/sim.8981  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42. Moher D, Schulz KF, Simera I, Altman DG. Guidance for developers of health research reporting guidelines. PLoS Med 2010;7:e1000217. 10.1371/journal.pmed.1000217  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43. Stewart LA, Clarke M, Rovers M, et al. PRISMA-IPD Development Group . Preferred Reporting Items for Systematic Review and Meta-Analyses of individual participant data: the PRISMA-IPD Statement. JAMA 2015;313:1657-65. 10.1001/jama.2015.3656  [DOI] [PubMed] [Google Scholar]
  • 44. Vandenbroucke JP, von Elm E, Altman DG, et al. STROBE Initiative . Strengthening the Reporting of Observational Studies in Epidemiology (STROBE): explanation and elaboration. PLoS Med 2007;4:e297. 10.1371/journal.pmed.0040297  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45. Benchimol EI, Smeeth L, Guttmann A, et al. RECORD Working Committee . The REporting of studies Conducted using Observational Routinely-collected health Data (RECORD) statement. PLoS Med 2015;12:e1001885. 10.1371/journal.pmed.1001885  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46. Moons KGM, Wolff RF, Riley RD, et al. PROBAST: A Tool to Assess Risk of Bias and Applicability of Prediction Model Studies: Explanation and Elaboration. Ann Intern Med 2019;170:W1-33. 10.7326/M18-1377  [DOI] [PubMed] [Google Scholar]
  • 47. Wolff RF, Moons KGM, Riley RD, et al. PROBAST Group† . PROBAST: A Tool to Assess the Risk of Bias and Applicability of Prediction Model Studies. Ann Intern Med 2019;170:51-8. 10.7326/M18-1376  [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Web appendix: Supplementary information

debt071018.ww.pdf (457.5KB, pdf)

Supplementary table: TRIPOD-Cluster checklist

debt071018.wt.pdf (82.3KB, pdf)

Articles from The BMJ are provided here courtesy of BMJ Publishing Group

RESOURCES