Abstract
Background
This study was designed to review the methodology and reporting of gastric cancer prognostic models and identify potential problems in model development.
Methods
This systematic review was conducted following the CHARMS checklist. MEDLINE and EMBASE were searched. Information on patient characteristics, methodological details, and models' performance was extracted. Descriptive statistics was used to summarize the methodological and reporting quality.
Results
In total, 101 model developments and 32 external validations were included. The median (range) of training sample size, number of death, and number of final predictors were 360 (29 to 15320), 193 (14 to 9560), and 5 (2 to 53), respectively. Ninety-one models were developed from routine clinical data. Statistical assumptions were reported to be checked in only nine models. Most model developments (94/101) used complete-case analysis. Discrimination and calibration were not reported in 33 and 55 models, respectively. The majority of models (81/101) have never been externally validated. None of the models have been evaluated regarding clinical impact.
Conclusions
Many prognostic models have been developed, but their usefulness in clinical practice remains uncertain due to methodological shortcomings, insufficient reporting, and lack of external validation and impact studies.
Impact
Future research should improve methodological and reporting quality and emphasize more on external validation and impact assessment.
1. Introduction
Although gastric cancer has been decreasing in terms of both incidence and mortality in most developed countries in recent decades, it still causes substantial disease burden [1]. Over 90% of people with early stage gastric cancer survive for five years or longer after surgical treatment [2, 3], whereas the prognosis of those with advanced gastric cancer is poor, with a 5-year survival rate of about 20% in stage III and 7% in stage IV patients, respectively [4].
Risk stratification is important in informing treatment decision, resource allocation, and patient recruitment in clinical trials [5]. The tumor-node-metastasis (TNM) staging system is widely used for risk stratification [4]. Nevertheless, even patients with the same TNM stage may have significantly different responses to treatment and clinical outcomes [6], suggesting that more accurate stratification could be beneficial [7]. Previous studies have identified numerous other prognostic factors for gastric cancer, which can be broadly divided into four categories: patient-, tumor status-, biomarker-, and treatment-related factors [8–10]. As no single prognostic factor suffices for satisfactory risk stratification, much interest has been raised in developing multivariable prognostic models, which quantitatively combine two or more prognostic factors [11, 12].
The American Joint Committee on Cancer has increasingly recognized the importance of incorporating prognostic models into practice to achieve personalized cancer management [13]. However, despite plenty of prognostic models published in the literature, very few have been adopted in clinical use. We carried out this systematic review to summarize the characteristics of existing models for predicting overall survival of patients with primary gastric cancer, with an emphasis on identifying potential problems in model development and validation and informing future research.
2. Methods
This systematic review was conducted following the CHARMS checklist [14], which was developed to guide data extraction and critical appraisal in systematic reviews of prediction model studies.
2.1. Eligibility Criteria
We included primary studies that reported the development and/or validation of prognostic models predicting overall survival of patients with primary gastric cancer. A prognostic model was defined as a combination of at least two prognostic factors, based on multivariable analysis, to estimate individual risk of a specific outcome, presented as regression formula, nomogram, or in a simplified form, such as risk score [15–17]. We only included prognostic models for predicting overall survival or all-cause death, excluding other outcomes, such as progression-free survival after treatment or disease-specific survival.
We excluded studies if (1) they enrolled patients with other types of cancer and the information on gastric cancer model could not be separately extracted; (2) they used short-term mortality (for example, death within 30 days after surgery) as the outcome; or (3) they validated prognostic models that were not initially developed for gastric cancer patients.
2.2. Literature Search
We searched MEDLINE and EMBASE to identify all relevant studies from their inceptions through 30 May 2018, using the following three groups of terms: (1) gastric tumor∗ OR gastric tumour∗ OR gastric cancer∗ OR gastric neoplasm∗ OR gastric carcinoma∗ OR stomach tumor∗ OR stomach tumour∗ OR stomach cancer∗ OR stomach neoplasm∗ OR stomach carcinoma∗ OR Siewert OR esophagogastric OR EGJ, and (2) prognos∗ OR survival OR death OR mortality, and (3) scor∗ OR model∗ OR index∗ OR nomogram∗ OR rule∗ OR predict∗ OR indices OR formula∗ OR equation∗ OR algorithm∗ [18, 19]. The search terms were limited to title and abstract. We also manually checked the reference lists of eligible studies to identify extra relevant studies.
2.3. Study Selection
After excluding duplicates, we screened all titles and abstracts to identify potentially eligible studies and then retrieved their full texts for further examination. Final eligibility was confirmed by two authors (QF and ZYY). Discrepancy was resolved by discussion with a third author (JLT).
2.4. Data Extraction and Quality Assessment
The data extraction form was designed according to the CHARMS checklist [14], supplemented with other items obtained from methodological guidance studies and previous systematic reviews [15, 16]. Briefly, the following information was extracted for each model development and external validation: publication year, country, data source, patient characteristics, length of follow-up, outcome, candidate predictors, training sample size, number of deaths, missing data, model development/validation methods, final predictors, predictive performance, and presentation formats.
In this study, candidate predictor refers to the potential predictors (and their functional forms, if any) that are selected to be examined in multivariable analysis but might or might not be included in final model. Final predictor refers to the predictors that are included in final models. Event per variable (EPV) is the ratio between the number of events and the number of candidate predictors, which is a rule of thumb to empirically evaluate the power of regression analysis, with a value of 10 or higher recommended to avoid potential overfitting [20, 21]. If one study included multiple model developments and validations, we extracted relevant information for each model development and validation separately. Data extraction was undertaken by two authors (QF and ZYY), and any uncertainty in data extraction was resolved by discussion with a third author (JLT).
We used a preliminary version of the Prediction Model Study Risk Of Bias Assessment Tool (PROBAST) to evaluate the methodological quality of each model development and validation [22]. This tool evaluated the levels of risk of bias in five domains: participant selection, definition and measurement of predictors, definition and measurement of outcome, sample size and participant flow, and statistical analysis. Each domain was rated as high, low, or unclear risk of bias. Overall judgment of risk of bias was derived from the judgments on all domains: low risk if all domains had low risk of bias, high risk if any domain had high risk of bias, otherwise unclear risk.
2.5. Statistical Analysis
We mainly used descriptive statistics to summarize the characteristics of model developments and validations. All final predictors were assigned into one of the four categories: patient, tumor status, biomarker, and treatment. We counted the frequency of each final predictor being included in models. We compared models that have been externally validated with those that have not, regarding their characteristics (training sample size, number of events, number of final predictors, EPV, c statistic, etc.).
3. Results
In total, 16334 citations were identified and 99 eligible publications (Supplementary ) were included in this review (Figure 1). Of the 99 publications, 75 performed model development only, 9 performed external validation only, and 15 performed both model development and external validation. One-hundred and one distinct models were extracted from 90 studies. Thirty-two external validations for 20 models were extracted from 24 studies.
3.1. Model Development
3.1.1. Basic Characteristics of Model Development
Characteristics of the 101 model developments are summarized in Table 1. Three models were published before 2000, and the number has rapidly risen by 2- and 30-fold during 2001–2010 and 2011–2018, respectively (Figure 2). Most models (76/101) originated from East Asian populations, which was not surprising, given the fact that this region has the highest incidence and prevalence of gastric cancer.
Table 1.
Model developments (n = 101) | |
---|---|
Study characteristics | |
Publication year | |
Before 2000 | 3 |
2001–2010 | 7 |
2011–2018 | 91 |
Study location | |
East Asia (China/Japan/Korea) | 76 |
Non-Asian | 25 |
Data source | |
Clinical data/retrospective cohort | 91 |
Prospective cohort | 7 |
Randomized controlled trial | 3 |
Patient characteristics | |
Male% (4/101 missing) | 67.6 (30.9, 80.3)a |
Age (5/101 missing) | |
Median (min, max) of mean | 60.0 (51.0, 70.0)a |
Tumor TNM stage | |
All | 46 |
I–III | 36 |
IV | 17 |
No information | 2 |
Gastrectomy | |
No restriction | 28 |
Only patients with gastrectomy | 71 |
Only patients without gastrectomy | 2 |
Model development | |
Sample size (training set) (14/101 missing) | 360 (29, 15320)a |
Number of events | 193 (14, 9560)a |
Event per variable (18/101 missing) | 25.1 (0.2, 1481.3)a |
Length of follow-up (month) (53/101 missing) | 44.0 (6.7, 111.6)a |
Start of outcome follow-up | |
From diagnosis | 3 |
From surgery | 49 |
From other time pointsb | 15 |
Unclear | 34 |
Candidate selection methods | |
Prespecification | 30 |
Univariable analysis | 63 |
Prespecification + univariable analysis | 5 |
Unclear | 3 |
Statistical model | |
Cox proportional hazard regression | 90 |
Othersc | 11 |
Final predictor selection | |
Full model | 10 |
Stepwise (including forward and backward) | 68 |
Unclear | 23 |
Statistical assumptions ever checked | 9 |
Number of final predictors | 5 (2, 53)a |
Formats of presentations | |
Score | 35 |
Nomogram | 47 |
Equation | 9 |
Others (decision tree and neural network) | 4 |
No | 6 |
Predictive performance | |
Discrimination | |
AUC/c statistic | 67 |
Others | 1 |
No | 33 |
Calibration | |
Calibration plot | 45 |
Hosmer–Lemeshow test | 3 |
No | 55 |
Model validation | |
Internal | 30 |
External | 21 |
No | 54 |
aMedian (min, max). bInitiation of chemotherapy (n = 10), metastasis (n = 3), and randomization (n = 2). cCART, Cox Lasso, discrimination analysis, Weibull model, neural network, and logistic model. AUC: area under curve.
Patient characteristics varied substantially across studies in terms of age, sex, tumor status, and treatment (Table 1). The median proportion of male patients was 67.8% (range 30.9% to 80.3%), and the median age was 60 years (range 51 to 70). Thirty-six models were developed from gastric cancer patients at TNM stages I–III only and 17 models from TNM stage IV only. Most models (71/101) recruited only patients who had received surgery.
3.1.2. Summary of Model Development Methods
Most models (91/101) were developed by retrospective cohort studies based on routine clinical data, which were not collected for the purpose of model development. To deal with missing data, which is a common problem with routine clinical data, seven models adopted the multiple imputation approach, while the remaining 94 models conducted complete-case analysis. The medians of total sample size and number of events included in analysis were 360 (range 29 to 15320) and 193 (range 14 to 9560), respectively. The starting point of follow-up for overall survival varied across models. Seven studies did not report their candidate predictors clearly. EPV can be estimated in 83 model developments, with the median of 25.1 (range 0.2 to 1481.3). A favored EPV (>10) was achieved in 64 model developments.
As for selection of candidate predictors, 63 models used univariable analysis, 30 models prespecified candidate predictors based on clinical knowledge, five models employed a combination of the two, and the other three models did not specify this issue clearly. Various statistical models were used for prognostic model development, with Cox proportional hazard model being the most popular one (used in 90 models). Sixty-eight models used a stepwise approach in multivariable analyses to select final predictors. Statistical assumptions of the methods were examined and reported in only nine studies.
The median number of final predictors was 5 (range 2 to 10). In total, 180 different predictors were included, of which 21 were patient-related, 34 tumor-related, 116 biomarkers, and 9 treatment-related (Table 2; more details in the Supplementary ). The most consistent predictors for overall survival (included by more than 10 models) were age at diagnosis, sex, lymph node involvement, metastasis, invasion depth, TNM stage, tumor size, tumor site, differentiation, and histologic type, all of which were patient- and tumor-related.
Table 2.
Category | Number of predictors | Number of predictors selected multiple times | Predictors selected multiple timesa |
---|---|---|---|
Patient | 21 | 9 | Age, sex, ethnicity, performance score, year of diagnosis, family history, smoking, residency, and addiction to opium |
| |||
Disease status | 34 | 21 | T stage, N stage, TNM stage, tumor site, tumor size, differentiation, metastasis, histologic type, Lauren type, LN ratio, lymphovascular invasion, bone metastasis, Borrmann type, liver metastasis, number of metastasis sites, lung metastasis, number of examined LN, metastasis LN, perineural invasion, LODDS, and TTP after chemotherapy |
| |||
Biomarker | 116 | 19 | CEA, NLR, ALP, albumin, bilirubin, CA199, Hb, CES1, IS, LDH, LNR:ART, lymphocyte count, MGAT5, mGPS, NPTM, platelet, sodium, TNFRSF11A, and WBC |
| |||
Treatment | 9 | 6 | Chemotherapy, gastrectomy, lymphedenectomy, resection margin, extent of resection, and radiotherapy |
aThe table lists only the predictors that have been included more than once. LN: lymph node. LODDS: log odds of positive LN. CEA: carcinoembryonic antigen. NLR: neutrophil/lymphocyte ratio. ALP: alkaline phosphatase. Hb: hemoglobin. MGAT5: β1, 6-N-acetylglucosaminyltransferase-5. mGPS: modified Glasgow Prognostic Score. CA199: cancer antigen 199. NPTM: number of positive tumor markers (cancer antigen 125, CA199, CEA). WBC: white blood cell. TTP: time to progression.
The models were mostly presented in simplified forms, such as risk score (35/101) and nomogram (47/101). For model performance, 33 and 55 models did not report discrimination and calibration, respectively. Among the studies reporting relevant information, the median c statistic for discrimination was 0.748 (range 0.627 to 0.961). Forty-two models were compared with TNM stage alone regarding c statistic value, and all models outperformed TNM stage, with a median increase in c statistic value of 0.050 (range 0.015 to 0.180).
3.2. External Validation
There were 32 external validations for 20 distinct models, with 22 of them reporting in the same study as the model development. The majority (81/101) of models developed have not been externally validated. Five models were externally validated more than once, and two models [23, 24] more than five times. The characteristics of training datasets and validation datasets were compared in 19 external validations. Five validations did not assess discrimination, and 24 did not assess calibration (Table 3). The median (range) of c statistic for discrimination was 0.770 (0.576 to 0.868). The difference in c statistic between development and validation ranged from −0.044 to 0.290 with a median of 0.029.
Table 3.
External validations (n = 32) | |
---|---|
Data source | |
Clinical | 27 |
Prospective cohort | 3 |
Randomized controlled trial | 2 |
Validated in | |
The original development study | 22 |
Independent study | 10 |
Sample size for validation | 610 (71, 26019)a |
Discrimination | |
AUC/c statistic | 25 |
Others | 2 |
No | 5 |
Calibration | |
Calibration plot | 6 |
Hosmer–Lemeshow test | 2 |
Calibration in large | 1 |
No | 24 |
Compared validation set with development set | 19 |
aMedian (min, max). AUC: area under curve.
3.3. Quality Assessment
The model developments had either high (97/101) or unclear (4/101) risk of bias, and all model validations had high risk of bias. Ninety-one developments and 31 validations had high risk of bias in participant selection, mainly due to retrospective data collection. Forty-six developments and six validations had high risk of bias in sample size and participant flow, mainly due to small sample size and inappropriate method of dealing with missing data. Eighty-three developments and 13 validations had high risk of bias in analysis, mainly due to inappropriate method dealing with continuous variable, lack of statistical assumption examination, lack of overfitting detection, and insufficient reporting of model performance (Supplementary ).
3.4. Comparison of Externally Validated Models with Not-Validated Models
When comparing development characteristics of externally validated models with not-validated models, we found that the validated models tended to have larger training sample size, bigger number of events, higher EPV, older age, and higher c statistic value, while the differences in number of final predictors seemed to be insignificant (Table 4). Multivariable logistic regression showed that models were more likely to be externally validated if they were developed with bigger training sample and higher c statistic.
Table 4.
Externally validated models (n = 20) mean (SD) | Not externally validated models (n = 81) mean (SD) | P value | |
---|---|---|---|
Training sample size | 3902.55 (5777.62) | 634.17 (926.30) | 0.021 |
Number of events | 2825.12 (4069.04) | 344.75 (613.35) | 0.028 |
Number of candidate predictors | 75.80 (204.53) | 12.83 (28.26) | 0.185 |
EPV | 364.21 (542.04) | 44.70 (82.97) | 0.033 |
Number of final predictors | 6.65 (3.44) | 5.94 (6.08) | 0.490 |
Length of follow-up (month) | 64.24 (29.65) | 43.76 (19.15) | 0.122 |
Age | 63.00 (4.99) | 59.87 (3.39) | 0.034 |
Male% | 64.92 (4.10) | 67.29 (6.54) | 0.053 |
c statistic | 0.80 (0.06) | 0.75 (0.07) | 0.042 |
EPV: event per variable.
4. Discussion
This systematic review identified 101 models predicting overall survival of gastric cancer patients, with 20 of them externally validated.
van den Boorn et al. published a systematic review [25] summarizing prediction models for esophageal and/or gastric cancer patients, but the present study substantially differed from it and has its own value. Firstly, this study focused on prognostic models designed for primary gastric cancer patients only, whereas van den Boorn et al. included models for both gastric and esophageal cancers. Secondly, we identified 40 more newly published models and provided a more comprehensive picture of their characteristics. Thirdly, van den Boorn et al. focused on models' performance and clinical application, but our study emphasized more on the methodology of model development and validation.
We observed substantial heterogeneity regarding patient types in model development. Many studies developed prognostic models for specific subgroups of gastric cancer patients (e.g., those with a certain tumor stage and those receiving certain treatment) to make their models unique from those developed by others. This strategy of patient restriction may limit the model's generalizability, increasing uncertainty when applying it to other types of patients. In addition, an underlying assumption of restricting a model development to specific patient subgroups is that there exists effect modification or interaction between the restriction variable(s) and the main prognostic factors of interest. However, most studies did not check this assumption.
We also identified common statistical shortcomings that may cause bias in model development. Firstly, most models were developed from routinely collected clinical data, in which missing data was common. Most models simply performed complete-case analysis by excluding the patients with missing data. However, complete-case analysis works well only when missing data occurs completely at random, which is rare in reality [26]. To address this issue, multiple imputation has been recommended [27, 28]. This method has been employed in prediction model studies of other diseases [29, 30] but has not been applied in gastric cancer until 2017 [31–33].
Secondly, univariable analysis was commonly used to select candidate predictors. However, this data-driven method has high risk of wrongly excluding a potentially significant variable or including a potentially insignificant variable when its association with the outcome is confounded by others [34, 35]. The bootstrap resampling method could be used to increase the stability of variable selection, by selecting variables with high inclusion frequency across multiple bootstrapping samples [36]. Moreover, variable selection should take into account clinical or biological knowledge and combine results of multivariable analysis with sensitivity analysis for cautious conclusion [34, 35].
Thirdly, the majority of studies did not examine the assumptions of the statistical models, such as hazard-proportionality for Cox regression and linearity assumption for continuous variables. The results of examination are important in selecting appropriate statistical models and determining predictors' functional forms [37, 38]. Cox regression is the most commonly used model for survival data, but the underlying proportional-hazard assumption was often violated, in which case parametric survival models could be considered [39, 40]. In addition, algorithms from machine learning (e.g., neural network) are less strict with assumptions and have been used more and more in prognostic model development.
Fourthly, detection of model overfitting was neglected in most studies. Overfitting is more likely to occur in studies with small sample size and many predictors, resulting in overestimation of risk in high-risk patients and underestimation low-risk ones [41]. This issue can be detected by internal validation with cross-validation or bootstrap resampling and handled with statistical methods, such as shrinkage and penalized regression [42, 43].
Underreporting is another common problem. Outcome definitions, variable selection method, assessment of discrimination, and calibration measures were not reported in 34%, 23%, 33%, and 55% of model developments, respectively. Because there is no standard method for model development and multiple feasible options exist at each step in model development, underreporting of methodological details may cause difficulty in assessing a model's internal and external validity. Future studies are suggested to follow relevant reporting guidelines such as the TRIPOD [44, 45].
In this study, we found 55 predictors that were included more than twice in models, 10 of which (age at diagnosis, sex, lymph node involvement, metastasis, invasion depth, TNM stage, tumor size, tumor site, differentiation, and histologic type) were included more than 10 times. This can be regarded as indirect evidence for their predictive power in gastric cancer prognosis. Direct evidence, i.e., magnitude of their association with the outcome, such as hazard ratio, can be found in previous systematic reviews and meta-analyses [46, 47]. Therefore, we suggest that future model development, if necessary, to build upon these existing evidence. Though a large number of biomarkers were studied, their frequency of being included in final models was very low (mostly once or twice).
Prognostic models can be used to inform patients of their prognosis and assist clinical decision-making. However, despite the much effort devoted into model development so far, very few prediction models other than the TNM stage system have been adopted in clinical practice. Apart from the problems discussed above, other reasons may include the complexity of those models as compared with TNM system and lack of external validation and clinical impact studies [48]. External validation evaluates a model's predictive performance in local setting and updates model if necessary. An impact study evaluates the effects of a prognostic model on clinical decision-making, behavioral change, subsequent health outcomes of individuals, and the cost-effectiveness of applying the model, with the optimal design being a clustered randomized controlled trial [49, 50]. The gap between prognostic model and clinical decision rule is another big concern. Prognostic models compute the probability of an event on a continuous scale or risk scores on an ordinal scale, whereas clinical decision is a binary choice regarding whether to use an intervention or not. Unfortunately, the translation of risk estimates derived from existing prognostic models to clinical decisions is much less investigated [50].
Therefore, future research should try to avoid repeatedly developing new models for similar predictive purposes with small sample size and high risk of bias. Instead, more emphasis should be put on improving methodological quality of model development, validating and updating models for use within their own setting [51], translating model prediction into clinical decision rules [50], and assessing the models' clinical impact [52, 53].
5. Conclusion
This systematic review identified 101 prognostic models for predicting overall survival of patients with gastric cancer, which were limited by high risk of bias, methodological shortcomings, insufficient reporting, and lack of external validation and clinical impact assessment. Future prognostic model research should pay more attention to their methodological and reporting quality, and more importantly, emphasized more on external validation and impact studies to assess the models' effectiveness in improving clinical outcomes.
Contributor Information
Zuyao Yang, Email: zyang@cuhk.edu.hk.
Jinling Tang, Email: jltang@cuhk.edu.hk.
Conflicts of Interest
The authors declare that they have no conflicts of interest.
Authors' Contributions
JLT and ZYY conceived the study. QF and ZYY did literature search and data extraction. QF analyzed the data. QF, ZYY, MTM, and SI interpreted the results. QF wrote the draft manuscript. All authors critically reviewed the manuscript.
Supplementary Materials
References
- 1.IARC. Population Fact Sheets: World. Lyon, France: IARC; 2018. http://gco.iarc.fr/today/fact-sheets-populations?population=900&sex=0. [Google Scholar]
- 2.Choi I. J., Lee J. H., Kim Y.-I., et al. Long-term outcome comparison of endoscopic resection and surgery in early gastric cancer meeting the absolute indication for endoscopic resection. Gastrointestinal Endoscopy. 2015;81(2):333.e1–341.e1. doi: 10.1016/j.gie.2014.07.047. [DOI] [PubMed] [Google Scholar]
- 3.Pyo J. H., Lee H., Min B.-H., et al. Corrigendum: long-term outcome of endoscopic resection vs. surgery for early gastric cancer: a non-inferiority-matched cohort study. American Journal of Gastroenterology. 2016;111(4):585–85. doi: 10.1038/ajg.2016.83. [DOI] [PubMed] [Google Scholar]
- 4.Washington K. 7th edition of the AJCC cancer staging manual: stomach. Annals of Surgical Oncology. 2010;17(12):3077–3079. doi: 10.1245/s10434-010-1362-z. [DOI] [PubMed] [Google Scholar]
- 5.Rugge M., Capelle L. G., Fassan M. Individual risk stratification of gastric cancer: evolving concepts and their impact on clinical practice. Best Practice & Research Clinical Gastroenterology. 2014;28(6):1043–1053. doi: 10.1016/j.bpg.2014.09.002. [DOI] [PubMed] [Google Scholar]
- 6.Qu J.-L., Qu X.-J., Li Z., et al. Prognostic model based on systemic inflammatory response and clinicopathological factors to predict outcome of patients with node-negative gastric cancer. PLoS One. 2015;10(6) doi: 10.1371/journal.pone.0128540.e0128540 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Asare E. A., Washington M. K., Gress D. M., Gershenwald J. E., Greene F. L. Improving the quality of cancer staging. CA: A Cancer Journal for Clinicians. 2015;65(4):261–263. doi: 10.3322/caac.21284. [DOI] [PubMed] [Google Scholar]
- 8.Allgayer H., Heiss M. M., Schildberg F. W. Prognostic factors in gastric cancer. British Journal of Surgery. 1997;84(12):1651–1664. doi: 10.1046/j.1365-2168.1997.00619.x. [DOI] [PubMed] [Google Scholar]
- 9.Mohri Y., Tanaka K., Ohi M., Yokoe T., Miki C., Kusunoki M. Prognostic significance of host-and tumor-related factors in patients with gastric cancer. World Journal of Surgery. 2010;34(2):285–290. doi: 10.1007/s00268-009-0302-1. [DOI] [PubMed] [Google Scholar]
- 10.Yasui W., Oue N., Aung P. P., Matsumura S., Shutoh M., Nakayama H. Molecular-pathological prognostic factors of gastric cancer: a review. Gastric Cancer. 2005;8(2):86–94. doi: 10.1007/s10120-005-0320-0. [DOI] [PubMed] [Google Scholar]
- 11.Edge S. B., Compton C. C. The American Joint Committee on Cancer: the 7th edition of the AJCC cancer staging manual and the future of TNM. Annals of Surgical Oncology. 2010;17(6):1471–1474. doi: 10.1245/s10434-010-0985-4. [DOI] [PubMed] [Google Scholar]
- 12.Riley R. D., Hayden J. A., Steyerberg E. W., et al. Prognosis research strategy (PROGRESS) 2: prognostic factor research. PLoS Medicine. 2013;10(2) doi: 10.1371/journal.pmed.1001380.e1001380 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Kattan M. W., Hess K. R., Amin M. B., et al. American Joint Committee on Cancer acceptance criteria for inclusion of risk models for individualized prognosis in the practice of precision medicine. CA: A Cancer Journal for Clinicians. 2016;66(5):370–374. doi: 10.3322/caac.21339. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Moons K. G. M., de Groot J. A. H., Bouwmeester W., et al. Critical appraisal and data extraction for systematic reviews of prediction modelling studies: the CHARMS checklist. PLoS Medicine. 2014;11(10) doi: 10.1371/journal.pmed.1001744.e1001744 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Mallett S., Royston P., Waters R., Dutton S., Altman D. G. Reporting performance of prognostic models in cancer: a review. BMC Medicine. 2010;8(1):p. 21. doi: 10.1186/1741-7015-8-21. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Mallett S., Royston P., Dutton S., Waters R., Altman D. G. Reporting methods in studies developing prognostic models in cancer: a review. BMC Medicine. 2010;8(1):p. 20. doi: 10.1186/1741-7015-8-20. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Damen J. A. A. G., Hooft L., Schuit E., et al. Prediction models for cardiovascular disease risk in the general population: systematic review. BMJ. 2016;353:p. i2416. doi: 10.1136/bmj.i2416. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Wilczynski N. L., Haynes R. B. Optimal search strategies for detecting clinically sound prognostic studies in EMBASE: an analytic survey. Journal of the American Medical Informatics Association. 2005;12(4):481–485. doi: 10.1197/jamia.M1752. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Wilczynski N. L., The Hedges Team, Haynes R. B. Developing optimal search strategies for detecting clinically sound prognostic studies in MEDLINE: an analytic survey. BMC Medicine. 2004;2(1):p. 23. doi: 10.1186/1741-7015-2-23. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Peduzzi P., Concato J., Feinstein A. R., Holford T. R. Importance of events per independent variable in proportional hazards regression analysis II. Accuracy and precision of regression estimates. Journal of Clinical Epidemiology. 1995;48(12):1503–1510. doi: 10.1016/0895-4356(95)00048-8. [DOI] [PubMed] [Google Scholar]
- 21.Concato J., Peduzzi P., Holford T. R., Feinstein A. R. Importance of events per independent variable in proportional hazards analysis I. Background, goals, and general strategy. Journal of Clinical Epidemiology. 1995;48(12):1495–1501. doi: 10.1016/0895-4356(95)00510-2. [DOI] [PubMed] [Google Scholar]
- 22.Wolff R. F., Moons K. G. M, Riley R. D., et al. Cochrane Database of Systematic Reviews. Cape Town, South Africa: The Global Evidence Summit; 2017. PROBAST: a risk-of-bias tool for prediction-modelling studies. [Google Scholar]
- 23.Woo Y., Son T., Song K., et al. A novel prediction model of prognosis after gastrectomy for gastric carcinoma. Annals of Surgery. 2016;264(1):114–120. doi: 10.1097/SLA.0000000000001523. [DOI] [PubMed] [Google Scholar]
- 24.Han D.-S., Suh Y.-S., Kong S.-H., et al. Nomogram predicting long-term survival after D2 gastrectomy for gastric cancer. Journal of Clinical Oncology. 2012;30(31):3834–3840. doi: 10.1200/JCO.2012.41.8343. [DOI] [PubMed] [Google Scholar]
- 25.van den Boorn H. G., Engelhardt E. G., van Kleef J., et al. Prediction models for patients with esophageal or gastric cancer: a systematic review and meta-analysis. PLoS One. 2018;13(2) doi: 10.1371/journal.pone.0192310.e0192310 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Schafer J. L., Graham J. W. Missing data: our view of the state of the art. Psychological Methods. 2002;7(2):147–177. doi: 10.1037/1082-989x.7.2.147. [DOI] [PubMed] [Google Scholar]
- 27.Bartlett J. W., Seaman S. R., White I. R., Carpenter J. R., Alzheimer’s Disease Neuroimaging Initiative Multiple imputation of covariates by fully conditional specification: accommodating the substantive model. Statistical Methods in Medical Research. 2015;24(4):462–487. doi: 10.1177/0962280214521348. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Marshall A., Altman D. G., Holder R. L., Royston P. Combining estimates of interest in prognostic modelling studies after multiple imputation: current practice and guidelines. BMC Medical Research Methodology. 2009;9(1) doi: 10.1186/1471-2288-9-57. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Janssen K. J. M., Vergouwe Y., Donders A. R. T., et al. Dealing with missing predictor values when applying clinical prediction models. Clinical Chemistry. 2009;55(5):994–1001. doi: 10.1373/clinchem.2008.115345. [DOI] [PubMed] [Google Scholar]
- 30.Sultan A. A., West J., Grainge M. J., et al. Development and validation of risk prediction model for venous thromboembolism in postpartum women: multinational cohort study. BMJ. 2016;355:p. i6253. doi: 10.1136/bmj.i6253. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Narita Y., Kadowaki S., Oze I., et al. Establishment and validation of prognostic nomograms in first-line metastatic gastric cancer patients. Journal of Gastrointestinal Oncology. 2018;9(1):52–63. doi: 10.21037/jgo.2017.11.08. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Wang Z.-X., Qiu M.-Z., Jiang Y.-M., Zhou Z.-W., Li G.-X., Xu R.-H. Comparison of prognostic nomograms based on different nodal staging systems in patients with resected gastric cancer. Journal of Cancer. 2017;8(6):950–958. doi: 10.7150/jca.17370. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Yuan S.-Q., Wu W.-J., Qiu M.-Z., et al. Development and validation of a nomogram to predict the benefit of adjuvant radiotherapy for patients with resected gastric cancer. Journal of Cancer. 2017;8(17):3498–3505. doi: 10.7150/jca.19879. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Heinze G., Dunkler D. Five myths about variable selection. Transplant International. 2017;30(1):6–10. doi: 10.1111/tri.12895. [DOI] [PubMed] [Google Scholar]
- 35.Sun G.-W., Shook T. L., Kay G. L. Inappropriate use of bivariable analysis to screen risk factors for use in multivariable analysis. Journal of Clinical Epidemiology. 1996;49(8):907–916. doi: 10.1016/0895-4356(96)00025-x. [DOI] [PubMed] [Google Scholar]
- 36.Austin P. C., Tu J. V. Bootstrap methods for developing predictive models. The American Statistician. 2004;58(2):131–137. doi: 10.1198/0003130043277. [DOI] [Google Scholar]
- 37.Harrell F. E., Jr., Lee K. L., Mark D. B. Multivariable prognostic models: issues in developing models, evaluating assumptions and adequacy, and measuring and reducing errors. Statistics in Medicine. 1996;15(4):361–387. doi: 10.1002/(SICI)1097-0258(19960229)15:4<361::AID-SIM168>3.0.CO;2-4. [DOI] [PubMed] [Google Scholar]
- 38.Sauerbrei W., Royston P., Binder H. Selection of important variables and determination of functional form for continuous predictors in multivariable model building. Statistics in Medicine. 2007;26(30):5512–5528. doi: 10.1002/sim.3148. [DOI] [PubMed] [Google Scholar]
- 39.Royston P., Lambert P. C. Flexible Parametric Survival Analysis Using Stata: Beyong the Cox Model. College Station, TX, USA: Stata Press; 2011. [Google Scholar]
- 40.Crowther M. J., Lambert P. C. A general framework for parametric survival analysis. Statistics in Medicine. 2014;33(30):5280–5297. doi: 10.1002/sim.6300. [DOI] [PubMed] [Google Scholar]
- 41.Pavlou M., Ambler G., Seaman S. R., et al. How to develop a more accurate risk prediction model when there are few events. BMJ. 2016;353:p. i3235. doi: 10.1136/bmj.i3235. [DOI] [PubMed] [Google Scholar]
- 42.Benner A., Zucknick M., Hielscher T., Ittrich C., Mansmann U. High-dimensional Cox models: the choice of penalty as part of the model building process. Biometrical Journal. 2010;52(1):50–69. doi: 10.1002/bimj.200900064. [DOI] [PubMed] [Google Scholar]
- 43.Pavlou M., Ambler G., Seaman S., De Iorio M., Omar R. Z. Review and evaluation of penalised regression methods for risk prediction in low-dimensional data with few events. Statistics in Medicine. 2016;35(7):1159–1177. doi: 10.1002/sim.6782. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Moons K. G. M., Altman D. G., Reitsma J. B., et al. Transparent reporting of a multivariable prediction model for individual prognosis or diagnosis (TRIPOD): explanation and elaboration. Annals of Internal Medicine. 2015;162(1):W1–W73. doi: 10.7326/M14-0698. [DOI] [PubMed] [Google Scholar]
- 45.Collins G. S., Reitsma J. B., Altman D. G., Moons K. G. M. Transparent reporting of a multivariable prediction model for individual prognosis or diagnosis (TRIPOD): the TRIPOD statement. British Journal of Surgery. 2015;102(3):148–158. doi: 10.1002/bjs.9736. [DOI] [PubMed] [Google Scholar]
- 46.Zu G., Zhang T., Li W., Sun Y., Zhang X. Impact of clinicopathological parameters on survival after multiorgan resection among patients with T4 gastric carcinoma: a systematic review and meta-analysis. Clinical and Translational Oncology. 2017;19(6):750–760. doi: 10.1007/s12094-016-1600-3. [DOI] [PubMed] [Google Scholar]
- 47.Zhou Y., Yu F., Wu L., Ye F., Zhang L., Li Y. Survival after gastrectomy in node-negative gastric cancer: a review and meta-analysis of prognostic factors. Medical Science Monitor. 2015;21:1911–1919. doi: 10.12659/msm.893856. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48.Adams S. T., Leveson S. H. Clinical prediction rules. BMJ. 2012;344(1):p. d8312. doi: 10.1136/bmj.d8312. [DOI] [PubMed] [Google Scholar]
- 49.Janssen K. J. M., Moons K. G. M., Kalkman C. J., Grobbee D. E., Vergouwe Y. Updating methods improved the performance of a clinical prediction model in new patients. Journal of Clinical Epidemiology. 2008;61(1):76–86. doi: 10.1016/j.jclinepi.2007.04.018. [DOI] [PubMed] [Google Scholar]
- 50.Reilly B. M., Evans A. T. Translating clinical research into clinical practice: impact of using prediction rules to make decisions. Annals of Internal Medicine. 2006;144(3):201–209. doi: 10.7326/0003-4819-144-3-200602070-00009. [DOI] [PubMed] [Google Scholar]
- 51.Toll D. B., Janssen K. J. M., Vergouwe Y., Moons K. G. M. Validation, updating and impact of clinical prediction rules: a review. Journal of Clinical Epidemiology. 2008;61(11):1085–1094. doi: 10.1016/j.jclinepi.2008.04.008. [DOI] [PubMed] [Google Scholar]
- 52.Collins G. S., Moons K. G. M. Comparing risk prediction models. BMJ. 2012;344(2) doi: 10.1136/bmj.e3186.e3186 [DOI] [PubMed] [Google Scholar]
- 53.Siontis G. C. M., Tzoulaki I., Siontis K. C., Ioannidis J. P. A. Comparisons of established risk prediction models for cardiovascular disease: systematic review. BMJ. 2012;344(1) doi: 10.1136/bmj.e3318.e3318 [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.