Abstract
Our work involves assessing whether new biomarkers might be useful for cervical-cancer screening across populations with different disease prevalences and biomarker distributions. When comparing across populations, we show that standard diagnostic accuracy statistics (predictive values, risk-differences, Youden’s index and Area Under the Curve (AUC)) can easily be misinterpreted. We introduce an intuitively simple statistic for a 2×2 table, Mean Risk Stratification (MRS): the average change in risk (pre-test vs. post-test) revealed for tested individuals. High MRS implies better risk separation achieved by testing. MRS has 3 key advantages for comparing test performance across populations with different disease prevalences and biomarker distributions. First, MRS demonstrates that conventional predictive values and the risk-difference do not measure risk-stratification because they do not account for test-positivity rates. Second, Youden’s index and AUC measure only multiplicative relative gains in risk-stratification: AUC=0.6 achieves only 20% of maximum risk-stratification (AUC=0.9 achieves 80%). Third, large relative gains in risk-stratification might not imply large absolute gains if disease is rare, demonstrating a “high-bar” to justify population-based screening for rare diseases such as cancer. We illustrate MRS by our experience comparing the performance of cervical-cancer screening tests in China vs. the USA. The test with the worst AUC=0.72 in China (visual inspection with acetic acid) provides twice the risk-stratification (i.e. MRS) of the test with best AUC=0.83 in the USA (human papillomavirus and Pap cotesting) because China has three times more cervical precancer/cancer. MRS could be routinely calculated to better understand the clinical/public-health implications of standard diagnostic accuracy statistics.
Keywords: AUC, Biomarkers, Cervical Cancer, Diagnostic Testing, HPV, Mean Risk Stratification, Number Needed to Test, Risk Prediction, ROC, Screening, Youden’s index
Introduction
After a new biomarker is convincingly associated with disease, a next question is to assess preliminarily, without enumerating costs/benefits/harms, whether the new biomarker is predictive enough to justify formal cost-effectiveness analyses that determine clinical/public-health use. For example, our work involves assessing whether new biomarkers might be useful for cervical-cancer screening in low/middle/high-resource populations, across which both disease prevalence and biomarker distributions vary substantially1.
Unfortunately, standard metrics reported for binary biomarkers provide at best indirect information about predictiveness for clinical/public-health use. The odds-ratio is a well-known poor measure of predictiveness2. When comparing two tests, it is uncommon for one test to have both higher sensitivity and specificity, or both higher positive predictive value (PPV) and negative predictive value (NPV). Two summary statistics, Youden’s Index 3 and the Area Under the Curve (AUC) statistic4, have been correctly criticized for not taking predictive values (i.e. absolute risks) into account, and for not permitting differential weighting of false-positives versus false-negatives5. AUC is the probability that someone with disease has higher risk than someone without disease, which requires only the risk ranks, not the absolute risks 4. Although AUC is used for continuous biomarkers, because the AUC for a binary test is a function of only Youden’s Index6, criticisms of Youden’s index also apply to AUC.
To better understand the implications of standard diagnostic accuracy metrics for clinical/public-health utility, we reinterpret standard metrics in light of a novel framework for quantifying risk-stratification. Although “risk-stratification” is a broadly used term, we define it as the ability of a test to separate those at high risk of disease from those at low risk7, allowing clinicians to intervene only for those that testing indicates are more likely to develop disease. We introduce two new broadly applicable, linked statistics that have proven useful in our epidemiologic work. We define mean risk-stratification (MRS) as the average change in risk of disease revealed by a test (post-test minus pre-test). We also define a complementary statistic, the Number Needed to Test (NNtest), which quantifies how many people require testing to identify one more disease case than would be identified by randomly selecting people for testing. All statistical details, including comparison to other statistics such as Decision Curves8, are in a preprint9. We have also used MRS to assess agreement between occupational exposure experts10.
MRS/NNtest have 3 advantages for comparing test performance across populations with different disease prevalences and biomarker distributions. First, PPV, NPV, and the risk-difference do not measure risk-stratification because they do not account for test-positivity rates, unlike MRS. That is, biomarkers that are rarely positive cannot provide much risk-stratification, regardless of how large the risk-difference is. Second, Youden’s index and AUC measure multiplicative relative gains in risk-stratification. However, large relative gains in risk-stratification might not imply large absolute gains if disease is rare. Third, the maximum MRS is limited by disease prevalence. Thus, little risk-stratification is possible for rare diseases such as cancer, demonstrating a “high-bar” to justify population-based screening. MRS/NNtest reinterpret the risk-difference, Youden’s index, and AUC, in the context of disease prevalence and biomarker distributions, which is crucial when considering test performance across populations. AUC cannot be used to rank tests between populations with different disease prevalence, and we show examples from our experience. Our webtool calculates MRS/NNtest (https://analysistools.nci.nih.gov/biomarkerTools/).
Background: Cervical-cancer screening tests
Human papillomavirus (HPV) causes almost all cervical cancer1. HPV-based screening is replacing cervical cytology (“Pap smears”), but many new tests are available or being developed11–12. For low/middle-income countries, a low-cost, but unreliable and inaccurate, test is visual inspection with acetic acid (VIA)13,14. Screen-positive women by any test are referred for definitive disease ascertainment by colposcopy. To expedite screening guidelines development15,16, we propose using MRS/NNtest to better identify potentially useful biomarkers.
To illustrate MRS/NNtest, we present data from 2 collaborations. Colleagues in China evaluated 3 tests (HPV testing, cervical cytology, and VIA) in an unscreened population of 30,371 women, to select a test as the basis for future nationwide screening17. Second, in support of US screening guidelines, we previously analyzed data on 331,818 women screened at Kaiser Permanente Northern California (KPNC) with cervical cytology, HPV testing, or both concurrently (“cotesting”)18.
Table 1 shows standard metrics for each test in China and KPNC. When comparing two tests in the same population, there is usually a tradeoff of sensitivity vs. specificity, or PPV vs. cNPV. This dilemma makes it hard to draw firm conclusions on the basis of any single statistic.
Table 1.
Population | Cervical screening testing modality | Odds ratio | Test positivity | PPV | cNPV | Sens | Spec | LR+ | LR− | Youden’s Index | AUC | MRS | NNtest |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Unscreened population in China | HPV | 206 | 17.3% | 9.1% | 0.05% | 98% | 84.0% | 6 | 0.03 | 82% | 91% | 2.60% | 77 |
Pap | 108 | 16.7% | 9.4% | 0.10% | 95% | 84.6% | 6 | 0.06 | 80% | 90% | 2.60% | 77 | |
VIA | 11 | 10.7% | 8.3% | 0.83% | 55% | 90.1% | 5 | 0.50 | 45% | 72% | 1.43% | 140 | |
Screened population KPNC (USA) | HPV/Pap Cotesting | 37 | 7.5% | 5.4% | 0.16% | 74% | 92.9% | 10 | 0.28 | 67% | 83% | 0.73% | 275 |
HPV | 47 | 5.1% | 7.6% | 0.17% | 70% | 95.3% | 15 | 0.31 | 65% | 83% | 0.71% | 280 | |
Pap | 13 | 3.8% | 4.7% | 0.36% | 34% | 96.3% | 9 | 0.69 | 30% | 65% | 0.32% | 633 |
Methods: Mean Risk Stratification (MRS) and Number Needed to Test (NNtest)
In the absence of test results or other pre-test information, each individual can only be assigned as a best guess the same population-average risk π=P(D+). After taking a test, people learn how their predicted individual risks differ from population-average. Two outcomes are possible:
With probability p=P(M+), the test is positive. The person’s risk increases from π=P(D+) to Positive Predictive Value (PPV=P(D+|M+)), an increase of PPV − π.
With probability P(M−)=1− p, the test is negative. The person’s risk decreases from π=P(D+) to complement of Negative Predictive Value: cNPV=1−NPV=P(D+|M−). The person’s risk decreases by π − cNPV.
Mean Risk Stratification (MRS) is a weighted average of the increase in risk among those who test positive and the decrease in risk among those who test negative:
(1) |
MRS is the average difference between predicted post-test individual risk and population-average (pre-test) risk. Simply, MRS is the average change in risk revealed by the test. Useful tests reveal that post-test risk is far enough from pre-test risk to justify changing management.
MRS also measures association of disease and marker. Denoting P(D+,M+) as a joint probability, and because the two terms on the right-hand side of MRS (1) are equal9, thus
(2) |
MRS measures association as twice the covariance of disease and marker. MRS is zero if and only if disease and marker are not associated. MRS is between −0.5 and +0.5.9
An inverse expression of MRS is also useful. A test is only useful if it is substantially better than randomly identifying cases by sheer luck, for which the joint probability P(D+,M+) would equal P(D+)P(M+). The Number Needed to Test (NNtest), to identify 1 more disease case than by randomly selecting individuals for definitive disease ascertainment is
Thus, NNtest=2/MRS. Although MRS and NNtest convey the same information, they are on different scales. MRS is on the scale of absolute risk, allowing comparison to absolute risks of other diseases. NNtest presents test performance in terms of numbers of people tested, which helps with considering the benefits, harms, and costs associated with the test and interventions. NNtest is not the Number Needed to Treat (NNT), which is the reciprocal of the risk-difference: NNT=(PPV−cNPV)−1. NNT conditions on test results, which is useful when exposures can be assigned, as in clinical trials. For deciding whether to screen a population, NNT is less relevant because test results cannot be assigned and the number testing positive is initially unknown.
For example, HPV testing in an unscreened population in China17 changed the risk of cervical precancer/cancer on average by MRS=2.6% (Table 1), which seems important because that is greater than precancer/cancer prevalence of 1.6%. In contrast, at KPNC, a heavily screened USA population, HPV testing had MRS=0.71%. A KPNC woman’s risk changed by only 0.71% on average by taking an HPV test, because cervical precancer/cancer prevalence is only 0.55% at KPNC. Similarly, only NNtest=140 Chinese women are required for VIA to identify 1 more precancer/cancer than would referring random women to colposcopy, versus the much higher NNtest=280 for HPV testing at KPNC, because disease is much more prevalent in China than at KPNC.
MRS/NNtest convey that cervical screening is less informative and efficient in the KPNC USA vs. China populations, due to previous screening and treatment reducing disease prevalence in the USA. Broadly, MRS/NNtest demonstrate inherent limits of screening for rare diseases like cancer19: although the few individuals testing positive may have large change in risk, the average change in risk following testing all individuals (the big majority of whom test negative) is modest. Hundreds may require screening to identify only one more disease case.
MRS has a simple formula for the standard cross-sectional population 2×2 table (n.b. not sampled case-control counts). For cell frequencies a=P(D+,M+), b=P(D+,M−), c=P(D−,M+), d=P(D−,M−), and inserting P(D+)=a+b and P(M+)=a+c into MRS equation (2):
(3) |
MRS is easy to remember as twice the cross-product difference of the frequencies in the interior of the 2×2 table for a cross-sectional population or unstratified random sample. In contrast, the odds-ratio is the cross-product ratio ad/bc. The odds-ratio is dimensionless, but MRS is on the scale of risk, and NNtest represents a number of people. In particular, if a,b,c,d represent cell-counts (not frequencies), with n being their sum, then (see Table 2):
n2 does not cancel in MRS/NNtest as it does in the odds ratio (i.e. (ad/n2) / (bc/n2) = ad/bc).
Table 2.
Test Result
|
|||
---|---|---|---|
Positive | Negative | Total | |
|
|||
Diseased | a | b | a+b |
Not Diseased | c | d | c+d |
Total | a+c | b+d | n=a+b+c+d |
Odds Ratio: | OR = ad/bc | ||
Mean Risk Stratification: | MRS = 2(ad−bc)/n2 | ||
Number Needed to Test: | NNtest = n2/(ad−bc) |
MRS is also the numerators of Pearson’s correlation, the Mantel-Haenszel test20, and Kappa10 (see eAppendix for relation to other statistics). Our preprint9 provides all statistical details, including variances, confidence intervals, and hypothesis testing.
Results: Implications of MRS and NNtest for evaluating diagnostic tests
1. Rarely positive tests with good PPV and cNPV provide little risk-stratification
Intuitively, but incorrectly, risk-stratification seems to depend only on how far the absolute risks, the PPV and cNPV, spread apart: the risk-difference RD=PPV − cNPV. Substitute into the MRS equation (1) P(D+) = PPV×P(M+) + cNPV×P(M−), and denoting p=P(M+), yields:
Risk-stratification depends on not only the spread between absolute risks (risk-difference RD), but crucially, also on test-positivity p. Thus, a rarely positive test, no matter how big the risk-difference, provides little risk-stratification.
For example, consider the hypothetical use of cervical cytology (contrary to usual practice to make a point), dichotomized simply as a cytologic result of cancer or not cancer (“not cancer” includes all precancerous and negative cytology results). At KPNC, the 5-year PPV for finding a precancer or cancer, following a cytology result of cancer, is 84%, the cNPV is 0.519%, and the overall precancer/cancer prevalence is 0.524%21. The risk-difference (RD) of 84%–0.519% = 83.5% is enormous, suggesting high risk-stratification. However, the probability of having a cytology result of cancer is only 0.006%. Consequently, only with probability 0.006% does a woman get the dramatic risk increase PPV−P(D+) = 84%−0.524% = 83.5%. Almost all the time (99.994%), she has a trivial risk decrease P(D+) −cNPV = 0.524%−0.519% = 0.005%. The MRS is 0.01%: only 1 extra precancer/cancer on average will be found over 5 years per 10,000 women using this test versus random selection. The NNtest=20,086 is an enormous number of women to screen to improve upon randomly selecting women to screen. Thus, good risk-stratification requires both a large risk-difference and a test that is positive sufficiently often.
Figure 1 plots the relationship of MRS/NNtest to the risk-difference for 3 test positivity rates. When risk-difference is 1, the maximum MRS and minimum NNtest are achieved. The importance of test positivity is illustrated by noting that, the MRS/NNtest achieved for risk-difference of 1 is also obtained for a risk-difference of approximately only 0.1 when the test is 10 times as positive (dashed line). Thus, a perfect marker for a rarely positive test provides the same risk-stratification as a much weaker-associated marker that is 10 times as positive.
2. AUC and Youden’s index do not fully measure risk-stratification, because risk-stratification depends strongly on disease prevalence
MRS/NNtest can be calculated by combining disease prevalence with Youden’s index (J=Sens+Spec−1) or AUC for a binary marker. Writing each joint probability in the MRS equation (3) as a function of sensitivity (Sens) and specificity (Spec), their respective complements cSens=1−Sens and cSpec=1−Spec, and disease prevalence π =P(D+):
Although AUC is mostly used for continuous markers, for binary tests, the AUC is the average of sensitivity and specificity6, which relates AUC to Youden’s index as J = 2×AUC − 1. Thus MRS, in terms of binary AUC, is MRS=4π(1− π)×(AUC− 0.5).
The key point is that MRS and NNtest are functions of AUC/Youden’s index and disease prevalence. This leads to 4 key conclusions:
1. The risk-stratification implied by Youden’s index, AUC, or diagnostic likelihood ratios depends on disease prevalence. Good values of Youden’s index, AUC, or diagnostic likelihood ratios do not imply good risk-stratification if the disease is too rare
A test with good values of Youden’s index, AUC, or diagnostic likelihood ratios in a population with low disease prevalence may provide less risk-stratification than a test with worse Youden’s index, AUC, or diagnostic likelihood ratios in a population with higher disease prevalence.
In Table 1, cotesting at KPNC in the USA has a greater AUC than VIA in China (83% in KPNC vs. 72% in China) and also better diagnostic likelihood ratios (LR+=10 vs. 5 & LR−=0.28 vs. 0.50). However, VIA in China has twice the MRS (0.7% vs. 1.4%), and half the NNtest (275 vs. 140), of cotesting at KPNC. Cotesting is incontestably more accurate than VIA14, but VIA stratifies risk better in an easier setting (unscreened population) than cotesting stratifies risk in a more challenging setting (heavily screened population). Thus, conducting even VIA in China could identify more treatable precancer (and hence prevent more cervical-cancer) than cotesting at KPNC. This point is obscured by the AUC and diagnostic likelihood ratios, but revealed by MRS.
For fixed AUC, MRS and NNtest are dramatically better as disease becomes more prevalent (Figure 2). In contrast, increasing AUC only slowly increases MRS. Thus, a screening test considered infeasible for general-population screening, may be feasible for screening higher-risk subpopulations19.
2. Disease prevalence defines the maximal amount of risk-stratification that a test can provide. Rare diseases inherently have limited risk-stratification
Since J=1 is a perfect biomarker, the maximum MRS for a disease is 2π(1− π); minimum NNtest is {π(1− π)}−1. Maximal MRS is solely determined by disease prevalence π=P(D+). The rarer the disease, the less risk-stratification is possible. If the maximum risk-stratification is small, every test must inevitably have low risk-stratification This is a reality of general-population cancer screening: on average, screening tests for uncommon diseases such as cancer do not reveal much risk information to tested individuals, and thus, hundreds require screening to identify one more disease case than would be found by randomly selecting people to screen.
In Table 1, the same tests have greater MRS (and lower NNtest) in an unscreened higher-risk population (China; 1.6% disease prevalence) than a screened low-risk population (USA KPNC; 0.55% disease prevalence). The maximum possible MRS is 3 times greater in China vs. KPNC (MRS: 3.2% vs. 1.1%; minimum NNtest: 64 vs. 183, respectively). A test with a fixed Youden’s index (or AUC) should stratify risk better in populations with more disease, such as previously unscreened populations or populations enriched for high-risk people (such as referral populations), because the maximum possible risk-stratification is greater.
Figure 2 plots the relationship of MRS and NNtest to AUC for 3 uncommon disease prevalences. The importance of disease prevalence is illustrated by noting that, the maximum MRS/NNtest (achieved for AUC=1) is also obtained if AUC=0.55 when disease is 10 times as prevalent (dashed line); AUC=0.6 is required if disease is 5 times as prevalent. Thus, a perfect marker for a rare disease provides the same risk-stratification as a weakly associated marker for a disease that is 5 or 10 times as prevalent.
3. Youden’s index is the percent of maximum risk-stratification actually achieved by the test. This interprets the meaning of increases in AUC in terms of risk-stratification
Because MRS = 2π(1− π)×J, and maximum MRS is 2π(1− π), thus the ratio of MRS to the maximum MRS is Youden’s index (J). Therefore, Youden’s index, instead of measuring risk-stratification, measures the percent of maximum risk-stratification actually achieved.
This expression helps interpret the sometimes obscure meaning of increases in AUC. Since J = 2×AUC − 1, each 1% increase in AUC implies a 2% increase in MRS (or decrease in NNtest). Thus, MRS doubles from AUC=0.6 to 0.7. An AUC=0.6 is widely considered to be “modest”, and indeed, only 20% of the maximum MRS is achieved. An AUC=0.7 is widely considered “good”, but only 40% of the maximum MRS is achieved. An AUC=0.9 is required to achieve 80% of the maximum MRS.
Although a high Youden’s index and AUC imply that a high percentage of the maximum risk-stratification has been achieved, they may conceal that there is little risk-stratification if disease is rare. The top axis of Figure 2 shows how the percent of maximum MRS increases linearly with AUC. As disease becomes rarer, such as for prevalence of 1/10,000, there is little discernible increase in MRS as AUC increases.
Similarly, we can interpret the risk-stratification implied by an odds-ratio. The minimum odds-ratio required to achieve a Youden’s index has sensitivity equal to specificity. Table 3 shows the minimum odds-ratio required to achieve each fraction of the maximum MRS (the maximum odds-ratio is always infinity, when sensitivity or specificity are 100%). For example, a marker with odds-ratio of 3.4 can attain no higher than 30% of the maximum MRS. If specificity must be high (say 95%), as is often the case for screening for rare diseases such as cancer, then OR=3.4 attains no higher than 10% of the maximum MRS. Seemingly large odds-ratios do not suffice to imply large risk-stratification. At KPNC, the 3 tests with the lowest and similar MRS of about 0.3% have odds-ratios varying widely (13-36) (Table 1). Similarly, the 4 best tests in China have similar MRS (2.6%-2.8%), but odds-ratios vary widely (108-206). Odds-ratios can reveal little about risk-stratification.
Table 3.
% of maximum risk-stratification achieved | minimum odds-ratio required
|
|
---|---|---|
overall | fix specificity=95% | |
| ||
0% | 1.0 | 1.0 |
10% | 1.5 | 3.4 |
20% | 2.3 | 6.3 |
30% | 3.4 | 10.2 |
40% | 5.4 | 15.5 |
50% | 9.0 | 23.2 |
60% | 16.0 | 35.3 |
70% | 32.1 | 57.0 |
80% | 81.0 | 107.7 |
90% | 361.0 | 361.0 |
100% | Infinity | Infinity |
4. It is important to distinguish rankings of different tests referring to the same population or between populations. Either can be useful depending on the objective
MRS/NNtest and J/AUC will rank the risk-stratification capability of tests in the same order in the same population, because disease prevalence is fixed (see Table 1). In Figure 1, the MRS increases monotonically with AUC. Thus comparing the MRS/NNtest for 2 tests in the same population is equivalent to comparing their J/AUCs, which is useful for hypothesis testing9. However, when prevalence varies across populations, J/AUC do not necessarily even rank test risk-stratification correctly according to MRS/NNtest. For example, a test with AUC=1 has less risk-stratification than any test with AUC>0.55 for a disease 10 times as prevalent (Figure 2, dashed line). VIA in China (AUC=0.72) provided twice the risk-stratification (i.e. MRS) of cotesting in the USA (AUC=0.83) in the USA. Thus, it is important to supplement AUC with MRS/NNtest to compare the risk-stratification of a test across populations with different disease prevalences.
Discussion and Conclusions
To enhance our work on assessing whether new biomarkers might be useful for cervical-cancer screening in low-, middle- and high-resource populations, we introduced two linked metrics, MRS and NNtest, that quantify risk-stratification from diagnostic tests. MRS and NNtest have 3 key advantages for comparing test performance across populations with different disease prevalences and biomarker distributions.
First, MRS and NNtest demonstrate that there is little risk-stratification possible for rarely positive tests, regardless of how high the risk-difference is. Second, MRS and NNtest demonstrate that tests with high AUC or Youden’s index do not necessarily imply high risk-stratification for rare diseases. Justifying screening for rare diseases, such as cancer, requires a strong justification that includes both sufficient prevalence and test positivity. China has 3 times the precancer/cancer prevalence of KPNC in the USA, and all tests, including VIA, the weakest cervical-screening test, provide far more risk-stratification in China than at KPNC. As expected, prior effective secondary-prevention screening reduces the value of more screening.
Third, MRS and NNtest demonstrate that Youden’s index and AUC measure multiplicative relative gains in risk-stratification, which might not imply large absolute gains if disease is too rare. An AUC=0.6 achieves only 20% of maximum risk-stratification; AUC=0.9 is required to achieve 80%. This is in accord with findings that AUC appropriately casts ‘pessimism’ on the power of risk prediction22,23. Within a population, Youden’s index and AUC will rank tests for risk-stratification in the same order as MRS and NNtest, although rankings may well differ across populations with differing disease prevalence.
This paper introduces MRS and NNtest to epidemiology. For diseases like cancer, the MRS is small and NNtest is large, and more experience is necessary to develop an intuitive sense of sufficient MRS and NNtest in various settings. Research on estimating MRS/NNtest from different study designs and statistical models is necessary. We have extended MRS to continuous markers and risk-prediction models, have compared MRS to new statistics, such as decision curves8, and calculated MRS at different risk-thresholds to identify “optimally informative” thresholds9. We have also used MRS to assess agreement between occupational exposure experts10. MRS should be considered alongside metrics of test accuracy, such as likelihood ratios and AUC, to reveal their risk stratification implications.
Standard diagnostic accuracy statistics are easy to misinterpret when comparing test performance across populations with differing disease prevalence or biomarker distributions. The STARD guidelines24 require reporting sensitivity/specificity, predictive values, or AUC, but these are hard to compare across populations and none directly informs about risk-stratification. MRS and NNtest might be worth adding to studies of test performance to quantify risk-stratification, improve cross-population comparisons, and better understand the meaning of standard statistics. For example, it is immediately clear from MRS and NNtest why BRCA1/2 testing of all women, justified on the basis of high PPV and risk-difference25, would have little public-health value because of low mutation-prevalence (0.25%). However, MRS and NNtest would also immediately clarify that restricting testing to populations with higher mutation-prevalence, such as Ashkenazi Jews (2.5%), is more likely to have public-health value9. MRS and NNtest can help temper premature claims of clinical/public-health utility and also help justify future definitive studies that account for benefits/harms/costs to decide clinical/public-health usefulness. Our risk-stratification webtool is publicly available: https://analysistools.nci.nih.gov/biomarkerTools/.
Supplementary Material
Highlights.
We propose new diagnostic accuracy metrics for comparing tests across populations
Rare diseases inherently permit limited risk-stratification
Tests that are rarely positive do not provide much risk-stratification
High AUC does not imply good risk-stratification if the disease is too rare
Our new metrics support a “high-bar” to justify population screening for cancer
Acknowledgments
We acknowledge our late friend, mentor, and collaborator, Dr. Sholom Wacholder, for his comments on earlier drafts of this work. We thank Anil Chaturvedi and Holly Janes for comments on earlier versions of this paper. We thank our colleagues at Kaiser Permanente Northern California and in China for their collaboration. We thank Christine Fermo and Sue Pan for helping develop the MRS Webtool at: https://analysistools.nci.nih.gov/biomarkerTools/.
Role of the funding source: This research was funded by the Intramural Research Program of the National Institutes of Health/National Cancer Institute, which reviewed the final manuscript for publication.
Footnotes
Publisher's Disclaimer: This is a PDF file of an unedited manuscript that has been accepted for publication. As a service to our customers we are providing this early version of the manuscript. The manuscript will undergo copyediting, typesetting, and review of the resulting proof before it is published in its final citable form. Please note that during the production process errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain.
Conflicts of Interest: None reported
References
- 1.Schiffman M, Castle PE, Jeronimo J, Rodriguez AC, Wacholder S. Human papillomavirus and cervical cancer. Lancet. 2007;370:890–907. doi: 10.1016/S0140-6736(07)61416-0. [DOI] [PubMed] [Google Scholar]
- 2.Pepe MS, Janes H, Longton G, Leisenring W, Newcomb P. Limitations of the odds ratio in gauging the performance of a diagnostic, prognostic, or screening marker. Am J Epidemiol. 2004;159:882–90. doi: 10.1093/aje/kwh101. [DOI] [PubMed] [Google Scholar]
- 3.Youden WJ. Index for rating diagnostic tests. Cancer. 1950;3:32–5. doi: 10.1002/1097-0142(1950)3:1<32::aid-cncr2820030106>3.0.co;2-3. [DOI] [PubMed] [Google Scholar]
- 4.Hanley JA, McNeil BJ. The meaning and use of the area under a receiver operating characteristic (ROC) curve. Radiology. 1982;143:29–36. doi: 10.1148/radiology.143.1.7063747. [DOI] [PubMed] [Google Scholar]
- 5.Greenhouse SW, Cornfield J, Homburger F. The Youden index: letters to the editor. Cancer. 1950;3:1097–101. doi: 10.1002/1097-0142(1950)3:6<1097::aid-cncr2820030620>3.0.co;2-e. [DOI] [PubMed] [Google Scholar]
- 6.Cantor SB, Kattan MW. Determining the area under the ROC curve for a binary diagnostic test. Med Decis Making. 2000;20:468–70. doi: 10.1177/0272989X0002000410. [DOI] [PubMed] [Google Scholar]
- 7.Wentzensen N, Wacholder S. From differences in means between cases and controls to risk stratification: a business plan for biomarker development. Cancer Discov. 2013;3:148–57. doi: 10.1158/2159-8290.CD-12-0196. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Vickers AJ, Elkin EB. Decision curve analysis: a novel method for evaluating prediction models. Med Decis Making. 2006;26:565–74. doi: 10.1177/0272989X06295361. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Katki HA. Novel decision-theoretic and risk-stratification metrics of predictive performance: Application to deciding who should undergo genetic testing. arXiv 2017;arXiv:1711.05686 [stat.ME] [Google Scholar]
- 10.Dopart PJ, Ji BT, Xue S, et al. Evaluating differences in expert agreement between subgroups to identify where to prioritize use of multiple raters. Submitted. [Google Scholar]
- 11.Castle PE, Stoler MH, Wright TC, Jr, Sharma A, Wright TL, Behrens CM. Performance of carcinogenic human papillomavirus (HPV) testing and HPV16 or HPV18 genotyping for cervical cancer screening of women aged 25 years and older: a subanalysis of the ATHENA study. Lancet Oncol. 2011;12:880–90. doi: 10.1016/S1470-2045(11)70188-7. [DOI] [PubMed] [Google Scholar]
- 12.Wentzensen N, Fetterman B, Castle PE, et al. p16/Ki-67 Dual Stain Cytology for Detection of Cervical Precancer in HPV-Positive Women. J Natl Cancer Inst. 2015;107:djv257. doi: 10.1093/jnci/djv257. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Shastri SS, Mittra I, Mishra GA, et al. Effect of VIA screening by primary health workers: randomized controlled study in Mumbai, India. J Natl Cancer Inst. 2014;106:dju009. doi: 10.1093/jnci/dju009. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Gravitt PE, Paul P, Katki HA, et al. Effectiveness of VIA, Pap, and HPV DNA testing in a cervical cancer screening program in a peri-urban community in Andhra Pradesh, India. PloS one. 2010;5:e13711. doi: 10.1371/journal.pone.0013711. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Massad LS, Einstein MH, Huh WK, et al. 2012 updated consensus guidelines for the management of abnormal cervical cancer screening tests and cancer precursors. Obstetrics and gynecology. 2013;121:829–46. doi: 10.1097/AOG.0b013e3182883a34. [DOI] [PubMed] [Google Scholar]
- 16.Schiffman M, Wentzensen N, Khan MJ, et al. Preparing for the Next Round of ASCCP-Sponsored Cervical Screening and Management Guidelines. Journal of lower genital tract disease. 2017;21:87–90. doi: 10.1097/LGT.0000000000000300. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Zhao FH, Hu SY, Zhang Q, et al. Risk assessment to guide cervical screening strategies in a large chinese population. Int J Cancer. 2016 doi: 10.1002/ijc.30012. [DOI] [PubMed] [Google Scholar]
- 18.Katki HA, Kinney WK, Fetterman B, et al. Cervical cancer risk for women undergoing concurrent testing for human papillomavirus and cervical cytology: a population-based study in routine clinical practice. Lancet Oncol. 2011;12:663–72. doi: 10.1016/S1470-2045(11)70145-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Castle PE, Katki HA. Screening: A risk-based framework to decide who benefits from screening. Nat Rev Clin Oncol. 2016;13:531–2. doi: 10.1038/nrclinonc.2016.101. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Lachin JM. Biostatistical Methods: The Assessment of Relative Risks. New York: Wiley-Interscience; 2000. [Google Scholar]
- 21.Katki HA, Schiffman M, Castle PE, et al. Benchmarking CIN 3+ risk as the basis for incorporating HPV and Pap cotesting into cervical screening and management guidelines. Journal of lower genital tract disease. 2013;17:S28–35. doi: 10.1097/LGT.0b013e318285423c. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Baker SG, Schuit E, Steyerberg EW, et al. How to interpret a small increase in AUC with an additional risk prediction marker: decision analysis comes through. Stat Med. 2014;33:3946–59. doi: 10.1002/sim.6195. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Pencina MJ, D’Agostino RB, Massaro JM. Understanding increments in model performance metrics. Lifetime Data Anal. 2013;19:202–18. doi: 10.1007/s10985-012-9238-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Bossuyt PM, Reitsma JB, Bruns DE, et al. STARD 2015: an updated list of essential items for reporting diagnostic accuracy studies. BMJ. 2015;351:h5527. doi: 10.1136/bmj.h5527. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.King MC, Levy-Lahad E, Lahad A. Population-based screening for BRCA1 and BRCA2: 2014 Lasker Award. JAMA. 2014;312:1091–2. doi: 10.1001/jama.2014.12483. [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.