Abstract
Cancer overdiagnosis is frequently estimated using the excess incidence in a screened group relative to that in an unscreened group. However, conditions for unbiased estimation are poorly understood. We developed a mathematical framework to project the effects of screening on the incidence of relevant cancers—that is, cancers that would present clinically without screening. Screening advances the date of diagnosis for a fraction of preclinical relevant cancers. Which diagnoses are advanced and by how much depends on the preclinical detectable period, test sensitivity, and screening patterns. Using the model, we projected incidence in common trial designs and population settings and compared excess incidence with true overdiagnosis. In trials with no control arm screening, unbiased estimates are available using cumulative incidence if the screen arm stops screening and using annual incidence if the screen arm continues screening. In both designs, unbiased estimation requires waiting until screening stabilizes plus the maximum preclinical period. In continued-screen trials and population settings, excess cumulative incidence is persistently biased. We investigated this bias in published estimates from the European Randomized Study of Screening for Prostate Cancer after 9–13 years. In conclusion, no trial or population setting automatically permits unbiased estimation of overdiagnosis; sufficient follow-up and appropriate analysis remain crucial.
Keywords: bias, early detection of cancer, mass screening, mathematical model, overdiagnosis, randomized clinical trial
Research articles on overdiagnosis—the detection of disease that would not present clinically in the absence of screening—have proliferated in recent years. In many of these studies, investigators considered the higher observed incidence (hereafter referred to simply as incidence) of disease in the presence of screening as a proxy for overdiagnosis (1–5). In such “excess incidence” studies, it has been estimated that overdiagnosis accounts for 31% of all breast cancer cases in the United States (1), 59% of screen-detected prostate cancer cases in the European Randomized Study of Screening for Prostate Cancer (ERSPC) (6, 7), and 22% of the cases detected in the mammography arm of the Canadian National Breast Screening Study (2). Concerns have been raised about the reliability of these estimates (8–11), although the precise conditions required for valid estimation are not well understood.
Overdiagnosis studies conducted in the population setting are challenging because, once screening has started, it is impossible to observe what the incidence would have been in the absence of screening. Counterfactual background incidence must therefore be extrapolated from historical trends or imputed by other means. The background incidence is critical in the calculation of excess incidence, because excess incidence is calculated as the difference between the observed incidence under screening and the background incidence. Population-based excess incidence estimates without a verifiably accurate estimate of background incidence should therefore be interpreted with great caution, particularly those involving extrapolation over a lengthy time interval.
Given that it can be difficult to determine background incidence in the population setting, assessments of overdiagnosis using incidence data from randomized screening trials may seem more reliable. Unlike population studies, trials offer an opportunity to validly estimate background incidence using data from the control group, so long as randomization is successful and there is no screening in the control arm (contamination). Because of the availability of a control group and perhaps because screening trials are considered gold-standard sources of evidence, the reliability of overdiagnosis estimates from these trials has not been thoroughly investigated.
In the present study, we examined conditions for valid estimation of cancer overdiagnosis based on excess incidence in both trial and population settings. To do so, we extended an early conceptual model (12) to quantitatively link cancer natural history, test sensitivity, and receipt of tests with the effects of a new screening program on observed disease incidence. Using the model, we derived conditions for valid empirical estimates of the number of overdiagnosed cancers and illustrated these conditions in common trial designs and plausible population settings. We concluded by evaluating empirical estimates in the ERSPC. We offer general recommendations below.
METHODS
Overview
In this section, we describe a model of the effects of a screening test on disease incidence that we can use to replicate the changes in incidence in trials and population settings under screening and to investigate bias associated with excess-incidence estimates of overdiagnosis. In general, the introduction of a screening program has predictable effects (12) on the incidence of relevant cancers, that is, cancers that would present clinically in the absence of screening. Initially, this incidence increases as the program reaches into the pool of preclinical cancers not detectable without the test. How far it reaches into the pool and how many preclinical cancers it nets—and therefore how much incidence increases and for how long—depends on the size of the pool and the characteristics of the program. Over time, however, as testing patterns stabilize and the pool of newly detectable cancers is emptied, incidence eventually falls. We used the model to formalize this process and develop expressions for the incidence of relevant cancers under screening and overdiagnosis, allowing us to derive the waiting time needed for unbiased empirical estimation of overdiagnosis using annual or cumulative incidence.
The model
The model comprises 5 components: 1) the rate of onset of relevant cancers, 2) the distribution of preclinical detectable time periods after onset, 3) the episode sensitivity of a new screening test, 4) the receipt of screening tests, and 5) the fraction of screen-detected cancers that are overdiagnosed. Components 1 and 2 determine the size of the pool of relevant cancers that are detectable by screening. In the absence of trends in risk factors or clinical practice before screening begins, the rate of onset of relevant cancers (r) will equal the rate of cancer diagnosis. In the United States, the annual incidence of breast cancer in 1975, before mammography screening was widely adopted, was 105 per 100,000 women (13). The annual incidence of prostate cancer in 1985, before prostate-specific antigen screening was introduced, was 116 per 100,000 men (13). Although clinical practice patterns were not constant before screening for these cancers began (12, 14), these incidence rates provide first approximations of the rates of onset of relevant breast and prostate cancers. In general, a higher rate of onset creates a larger reservoir of preclinical cancers that can be detected by screening.
Because relevant cancers are destined to be clinically diagnosed within the lifetime of the patient, the longest possible preclinical period for these cancers is finite. Let [0, D] denote the range of these periods in years. The periods when cancers are detectable by screening are sometimes called “sojourn times” (15–17). A closely related idea is the lead time, or the time by which screening advances diagnosis; the longest lead time provides a conservative approximation to the longest preclinical period (D). Lead time distributions have been estimated for several cancers (15, 16, 18) and average 2–4 years for invasive breast cancers and 5–7 years for prostate cancers in the United States. A longer preclinical period permits screening to detect relevant cancers earlier and to advance diagnosis of cancers that would have presented further in the future.
Components 3 and 4 determine how quickly the latent pool is depleted. Episode sensitivity (19) p represents the probability of diagnosis due to the test among individuals with preclinical disease. A more sensitive test, a greater frequency of compliance with biopsy recommendations, and a more sensitive biopsy all increase the episode sensitivity, which in turn more quickly depletes the pool of preclinical cancers. Similarly, a greater number of tests drains the pool more quickly. In a randomized trial, receipt of tests reflects attendance at invited screens q. In a population setting, receipt of tests reflects dissemination into clinical practice. Note that only in the unrealistic situation in which a test is done continuously and has perfect sensitivity is sojourn time equal to lead time.
Component 5 determines the number of overdiagnosed cancers each year, which we express as a constant fraction b of screen detections. We make no distinction between overdiagnosed cancers that would have progressed to symptomatic presentation were it not for death from another cause, cancers that are indolent, or cancers that would have regressed spontaneously. These distinctions do not impact our results.
Using these model components, we can derive expressions for cancer incidence with and without screening. Let denote the number of new relevant cancers that develop in year y with preclinical period k years, so that a total of cancers develop in year y, of which are diagnosed in year y, are diagnosed in year y + 1, and so on, until are diagnosed in year y + D. Thus, in each year y before screening begins, cancers reach the end of their preclinical periods and are clinically diagnosed. Figure 1 shows how incidence and prevalence of relevant cancers at a point in time are composed of cancers that developed during or before this time.
Suppose an annual screening program is introduced in year y*. Assume, without loss of generality, 100% attendance by the population. After screening begins, the number of cancers that reach the end of their preclinical period without being detected by screening and are clinically diagnosed in year y is:
(1) |
where y − y* is the number of years since the screening program began. Here min{k, y − y*} represents the number of screening tests given when preclinical disease is present, each of which must be a false negative and thus occurs with probability 1 − p. This formulation assumes that test results are independent and sensitivity is constant; generalizations in which sensitivity depends on the preclinical period or with proximity to clinical diagnosis can be readily derived. The incidence of screen-detected cancers is:
(2) |
In words, screen-detected incidence in year y is comprised of cancers that had not yet reached the end of their preclinical periods, that were not detected by previous tests, and that were detected by the test in year y. Finally, let Oy = bSy denote the number of overdiagnosed cancers diagnosed in year y, so that total incidence in year y is:
(3) |
Example trial and population settings
Given suitable inputs, the model can project total incidence and its components across years. We demonstrate these projections in trial and population settings for episode sensitivity p = 0.5 and overdiagnosis fraction b = 0.25. For simplicity, the number of preclinical cancers is determined by assuming that relevant cancers develop at an annual rate of r = 100 per 100,000 individuals for a population with fixed size P = 50,000 or 100,000 and the preclinical period follows a discrete uniform distribution so that for all k and y. The starting year of screening y* can vary between the screen and control arm in each trial design and across subpopulations that adopt annual screening in each population setting.
The trial designs are as follows. 1) In a stop-screen trial, the screen arm receives 4 annual tests and then screening stops. 2) In a continued-screen trial, the screen arm receives annual tests indefinitely. 3) In a delayed-screen trial, the screen arm receives annual tests indefinitely, and the control arm receives annual tests indefinitely after a 4-year delay. Similar designs have been used to study breast or prostate cancer screening. We assume an attendance rate of q = 0.8 at each screen test and a maximum preclinical period of D = 6 years.
The population settings are as follows. 1) Screening disseminates over a ramping-up period in which cumulative percentages of the population that adopt annual screening are 5%, 15%, 30%, 45%, and 50% in years 2, 3, 4, 5, and 6, respectively, and D = 6 years. 2) Annual screening is adopted by 10%, 30%, and 50% of the population in years 2, 3, and 4, respectively, and D = 6 years. 3) Screening is as in setting 1 and D = 12 years.
In both trial and population settings, the model projects annual and cumulative incidence over specified time periods. In each setting, we compute excess incidence in screened groups relative to unscreened groups, compare this with the true number of overdiagnosed cancers, and identify the earliest time point at which the empirical estimate is unbiased.
RESULTS
The introduction of a new screening program induces a bulge in the incidence of relevant cancers relative to background incidence. General results about the height and width of the bulge are derived in the Appendix. To summarize, for a given rate of onset, the height of the bulge is determined by the episode sensitivity and the receipt of tests, and the end of the bulge is given by the first point that screening stabilizes plus the maximum preclinical period. Thus, even when accurate information about background incidence is available, an unbiased empirical estimate of the number of overdiagnosed cancers requires follow-up at least as long as the longest preclinical period once screening stabilizes.
Figure 2 shows annual and cumulative incidence in the screen and control arms for each trial design, with overdiagnosed cancers highlighted in shaded areas. In the stop-screen design, when participants in the screen arm stop undergoing screening, its annual incidence (Figure 2A) falls below control-arm incidence because relevant cancers that would have presented then have already been detected. The annual incidence in the screen arm then gradually rises back to control-arm incidence. Excess cumulative incidence (Figure 2D) first equals the number of overdiagnosed cancers at 10 years of follow-up, corresponding to the first year screening stops (year 4) plus the 6-year maximum preclinical period.
If the screen arm continues screening, excess annual incidence (Figure 2B) first equals the number of overdiagnosed cancers at 7 years of follow-up, corresponding to the 6-year maximum preclinical period after screening starts (year 1). However, in contrast with the stop-screen design, excess cumulative incidence always overstates overdiagnosis (Figure 2E).
If the control arm starts screening after a delay (Figure 2F), it no longer provides information about incidence in an unscreened group, and no unbiased empirical estimate of overdiagnosis is available. If the frequency of overdiagnosis is the same in the 2 arms, annual incidence in the control and screened groups converge after 11 years (Figure 2C), corresponding to the first year screening begins in the control arm (year 5) plus the 6-year maximum preclinical period.
Figure 3 shows annual incidence in the population settings partitioned into screen and clinical diagnoses and overdiagnoses (equation 3 above). Once screening stabilizes, the effects of screening on incidence resemble patterns in a continued-screen trial. The first point the bulge in relevant cancers returns to the background level, so that excess incidence provides an unbiased estimate of overdiagnosed cancers, occurs 5 + 6 = 11 (Figure 3A), 3 + 6 = 9 (Figure 3B), and 5 + 12 = 17 (Figure 3C) years from the start of screening (year 2), reflecting the time for screening uptake to stabilize plus the maximum preclinical period. In general, the first point that excess incidence provides an unbiased estimate of overdiagnosis, as well as the bias before this point, is a complicated function of cancer natural history, episode sensitivity, and screening uptake.
Although excess annual incidence in a continued-screen trial or population setting eventually provides an unbiased estimate of the number of overdiagnosed cancers in a given year, excess cumulative incidence does not yield an accurate result, even after many years. This is because screening is continually advancing the diagnosis of relevant cancers that would have been diagnosed in the future, and background incidence under the same follow-up has not yet caught up. This is shown in Table 1, in which we report excess cumulative incidence and the true incidence of overdiagnosis in the 3 population settings in Figure 3. In each setting, the selected study period ranges from the start of screening to the first point that an unbiased estimate is available using excess annual incidence. We found that, in each setting, excess cumulative incidence was nearly double the true incidence of overdiagnosed cancers over this period.
Table 1.
Population Settinga | Setting Characteristics |
Cumulative Incidence (Cases per 100,000) |
Overdiagnosis Estimate (Cases per 100,000) |
Error in Excess Incidence Estimate |
|||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Cumulative Population Starting Annual Screening by Year, % |
Maximum Preclinical Detectable Period, Years | Study Years | Without Screening | With Screening | Excess Incidenceb | True Incidence of Overdiagnosis | Absolutec | Relatived | |||||
2 | 3 | 4 | 5 | 6 | |||||||||
1 | 5 | 15 | 30 | 45 | 50 | 6 | 2–12 | 1,100 | 1,339.1 | 239.1 | 124.9 | 114.2 | 1.91 |
2 | 10 | 40 | 50 | 6 | 2–10 | 900 | 1,128.3 | 228.3 | 114.2 | 114.2 | 2.00 | ||
3 | 5 | 15 | 30 | 45 | 50 | 12 | 2–18 | 1,700 | 2,228.1 | 528.1 | 270.4 | 257.7 | 1.95 |
a Setting 1 involves slow dissemination of screening and a short maximum preclinical period; setting 2 involves fast dissemination of screening and a short preclinical period; and setting 3 involves slow dissemination of screening and a long maximum preclinical period. In each setting, relevant cancers develop at an annual rate of 100 per 100,000 individuals, episode sensitivity is 50%, and 25% of screen-detected cancers are overdiagnosed.
b Excess incidence is cumulative incidence with screening minus cumulative incidence without screening.
c Absolute error is excess incidence minus true incidence of overdiagnosis.
d Relative error is excess incidence divided by true incidence of overdiagnosis.
DISCUSSION
Using excess incidence to estimate the number of overdiagnosed cancers is intuitive and relatively common. The potential for this approach to be misleading has been noted (20–22), but formal conditions for valid estimates have not been established. Our examination shows that, even in an uncontaminated screening trial or a population setting with perfect knowledge about background incidence, excess incidence yields a biased estimate of overdiagnosed cancers except under fairly specific circumstances. In particular, excess incidence in a stop-screen trial requires using cumulative incidence, whereas a continued-screen trial or population setting requires using annual incidence. Either setting further requires that excess incidence be calculated only after screening patterns have stabilized plus the maximum preclinical period. This finding confirms and formalizes results from prior studies (11, 20, 23) regarding the importance of adequate follow-up in excess-incidence studies of overdiagnosis.
The model involves several simplifying assumptions that were useful for deriving these general conditions. First, we assumed a discrete uniform distribution for the preclinical periods and did not allow this distribution to change over time. Further, incidence and screening frequencies did not vary with age. These assumptions did not affect our main conclusions, but the absence of an age structure limited our ability to interrogate proposed adjustments to excess incidence that use age-specific results to reduce bias (e.g., 8, 9). Test sensitivity was assumed independent of prior tests and of time to clinical diagnosis, though this assumption can be generalized. We made no assumptions about whether overdiagnoses represented cancers that were progressive, indolent, or regressive; in certain contexts, a practitioner may wish to represent competing mortality explicitly, for example, to tease out information about the proportion of overdiagnosed cancers that are progressive.
In certain settings, the model can be used to evaluate existing studies of excess incidence for their likely validity and to explore the potential magnitude of bias. To illustrate, we consider the ERSPC, a continuous-screen trial with limited screening in the control arm (24). ERSPC investigators reported cumulative incidence differences between the screen and control arms of 34, 36, and 33 cancers per 1,000 men at 9, 11, and 13 years of follow-up, respectively (6, 25, 26), and used these to approximate overdiagnosis when calculating the number needed to detect to prevent 1 prostate cancer death. Figure 4 shows reported cumulative excess incidence (black dots) and corresponding model projections for each arm (solid and dashed lines). For specified values of the maximum preclinical period, we identified overdiagnosis frequencies that yielded projected incidence that was similar to observed incidence in the screen arm. As expected, longer maximum preclinical periods among relevant cancers required a lower overdiagnosis frequency. Under a 12-year maximum preclinical period (Figure 4A), a 40% overdiagnosis frequency was required, corresponding to 16, 19, and 21 overdiagnoses per 1,000 men, approximately half the published excess cases at 9, 11, and 13 years, respectively. Under a 20-year maximum preclinical period (Figure 4C), which is more consistent with prevailing knowledge about prostate cancer natural history (27), a 10% overdiagnosis frequency was required, implying even greater bias in published estimates. Because our model relies on simplifying assumptions, this examination of ERSPC estimates should only be interpreted as suggestive of the potential magnitude of bias in published results. In practice, reliable empirical estimates of overdiagnosis in this trial can only be obtained by calculating the excess using annual instead of cumulative incidence, with reliability increasing with increasing length of follow-up.
In contrast with the ERSPC, the use of cumulative incidence in the Canadian National Breast Screening Study is appropriate given its stop-screen design and 20 years of follow-up after the end of the screening given published estimates of the preclinical period for breast cancer (16). However, screening was not tracked after the intervention period. If the screen arm continued screening and/or the control arm started screening, the validity of any estimate becomes difficult to judge.
Similarly, our findings can be used to evaluate empirical estimates from population studies. For instance, in a study of mammography screening in the United States, Bleyer and Welch (1) compared observed breast cancer incidence with assumed background rates over 30 years and estimated that 31% of breast cancers were overdiagnosed in 2008. Although the assumed background rates have been criticized (28, 29), follow-up for this calculation after stabilization of mammography screening may be sufficient in principle for valid estimation of overdiagnosis in 2008. However, in the same study (1), the authors found 1.3 million excess cumulative breast cancers among women during 1978–2008. By using excess cumulative incidence from the introduction of screening, the estimated overdiagnoses frequency likely includes relevant in addition to overdiagnosed cancers.
Although we zeroed in on the issues of study design, calculation metric, and follow-up duration, we did not consider modifications that have been suggested to reduce bias associated with excess-incidence estimates when screening is restricted to specific age strata (30). These modifications include comparing incidence under screening with incidence in an older population (to account for lead time) (8) and considering incidence under screening that includes ages beyond the screening age range (to account for the compensatory drop in incidence that occurs among persons older than those who are offered screening) (23). In principle, our model can be extended to investigate these modifications, but this will require imposing an age structure on the modeled population.
In previous work, researchers have identified the need for sufficient follow-up before an unbiased estimate of overdiagnosis is available (20, 23). Duffy and Parmar (23) conducted a quantitative investigation into the follow-up needed in a hypothetical population setting, but we know of no investigation in trial settings. Rather, there seems to be a prevailing confidence in overdiagnosis results from trials due to the randomization of subjects to intervention groups. Our work indicates that this is misplaced. Alternative approaches have also been used to estimate overdiagnosis, including statistical (17, 31, 32), analytic (33), microsimulation (34), and decision analysis (35) models. Strengths and limitations of these approaches have been previously discussed (10, 36–38). Further research is needed to identify the conditions under which model-based estimates provide valid estimates of overdiagnosis.
In conclusion, we offer the following recommendations for estimating overdiagnosis using the excess incidence in a screened relative to an unscreened group. 1) Calculate the difference only after the initial years of screening uptake. Use annual incidence in a continued-screen trial and cumulative incidence in a stop-screen trial. Note that the wait time condition requires sufficient follow-up for all individuals, and the often-reported median follow-up in a trial can be substantially longer than the complete follow-up necessary for valid estimation. In a population setting, compelling information is needed about a trend in background incidence for an empirical estimate of overdiagnosis to be plausible. 2) Review estimates of cancer natural history to evaluate consistency of overdiagnosis estimates with prevailing understanding of natural history and lead-time distributions (37). The maximum lead time is a useful lower bound for the maximum preclinical period. 3) Recognize that a trial setting provides a valid comparison group provided that noncompliance and contamination are minimal. However, this does not automatically mean that overdiagnosis estimates from a trial are valid. In fact, even in the trial setting, sufficient follow-up and appropriate analytic methods are necessary to correctly interpret excess incidence as an estimate of the frequency of overdiagnosis.
ACKNOWLEDGMENTS
Author affiliations: Division of Public Health Sciences, Fred Hutchinson Cancer Research Center, Seattle, Washington (Roman Gulati, Ruth Etzioni); and Division of Cancer Control and Population Sciences, National Cancer Institute, Bethesda, Maryland (Eric J. Feuer).
This work was supported by the National Cancer Institute at the National Institutes of Health (grants R01 CA192402 and U01 CA199338).
The contents of this study are solely the responsibility of the authors and do not necessarily represent the official views of the National Cancer Institute or the National Institutes of Health.
Conflict of interest: none declared.
APPENDIX
Minimum Follow-Up for Unbiased Empirical Estimation of Overdiagnosis
Using the model, we determine the minimum follow-up necessary for excess incidence to yield an unbiased estimate of the number of overdiagnosed cancers. Because this is equivalent to the first time point at which relevant cancers return to the background level after screening begins, we determine the earliest year y′ (with y′ ≥ y*) in which the number of new cancers that develop and enter the preclinical pool (Ny′) equals the number of cancers detected and removed from the pool (Cy′ + Sy′). Symbolically, this condition is:
(1) |
(We assume here 100% attendance with no loss of generality.) Under a stationary distribution of preclinical periods, in which the same number of relevant cancers with a given preclinical period develop each year, the steady state is reached precisely when it is reached for each preclinical period k = 1, 2, …, D, that is, when
(2) |
(We can ignore k = 0 because the cancers that develop in year y′ are diagnosed in that year.) Under a stationary distribution of preclinical periods, the number of cancers with the same preclinical period k cancels from both sides, and equation 2 holds precisely when
(3) |
To see that this condition holds when y′ − y* ≥ D, first notice that when k = 1, it is trivial that 1 = (1 − p) + p. If this condition holds when k = n, then
(4) |
and so the condition holds by mathematical induction. In contrast, when 0 ≤ y′ − y* < D, 1 or more terms necessary for equality in equation 3 are missing from the right side, which implies that the number of cancers detected exceeds the number of new cancers that develop during this period. In other words, once screening begins, incidence of relevant cancers first increases above the background level and then eventually returns to this level after the maximum preclinical period has elapsed.
If screening disseminates into the population over time, it is necessary to wait until screening stabilizes across the population plus the maximum preclinical period. Before this point, the number of cancers being detected and removed from the population is still ramping up. Similarly, if test sensitivity improves over time, screening will reach progressively deeper into the pool of preclinical cancers, and incidence of relevant cancers will achieve a long-term steady state only after sensitivity has stabilized. In general, an unbiased estimate of overdiagnosis requires a new batch of cancers with the full range of preclinical periods to develop and be diagnosed under stable screening practices.
If the distribution of preclinical periods is not stationary, for example, because the population is open and includes entrance of individuals with different distributions of preclinical periods than the initial population or because there is a secular trend in factors associated with cancer risk, the situation is more complicated. The wait time necessary for the incidence of relevant cancers to achieve a steady state will depend on the effects of the new entrants on the pool of preclinical cancers. Similarly, if test sensitivity depends on the preclinical period or proximity to clinical diagnosis, dynamic effects on the necessary wait time will depend on the specific details of these relationships. Broadly speaking, however, time-varying effects of screening will tend to lengthen the wait time necessary for incidence of relevant cancers to return to the background level. Even in the simplest setting, convergence to background incidence is a complicated function of the rate of onset, the distribution of preclinical periods, test sensitivity, and receipt of tests.
REFERENCES
- 1.Bleyer A, Welch HG. Effect of three decades of screening mammography on breast-cancer incidence. N Engl J Med. 2012;36721:1998–2005. [DOI] [PubMed] [Google Scholar]
- 2.Miller AB, Wall C, Baines CJ et al. . Twenty five year follow-up for breast cancer incidence and mortality of the Canadian National Breast Screening Study: randomised screening trial. BMJ. 2014;348:g366. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Welch HG, Albertsen PC. Prostate cancer diagnosis and treatment after the introduction of prostate-specific antigen screening: 1986–2005. J Natl Cancer Inst. 2009;10119:1325–1329. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Kalager M, Adami H-O, Bretthauer M et al. . Overdiagnosis of invasive breast cancer due to mammography screening: results from the Norwegian screening program. Ann Intern Med. 2012;1567:491–499. [DOI] [PubMed] [Google Scholar]
- 5.Morrell S, Barratt A, Irwig L et al. . Estimates of overdiagnosis of invasive breast cancer associated with screening mammography. Cancer Causes Control. 2010;212:275–282. [DOI] [PubMed] [Google Scholar]
- 6.Schröder FH, Hugosson J, Roobol MJ et al. . Screening and prostate-cancer mortality in a randomized European study. N Engl J Med. 2009;36013:1320–1328. [DOI] [PubMed] [Google Scholar]
- 7.Welch HG, Black WC. Overdiagnosis in cancer. J Natl Cancer Inst. 2010;1029:605–613. [DOI] [PubMed] [Google Scholar]
- 8.Biesheuvel C, Barratt A, Howard K et al. . Effects of study methods and biases on estimates of invasive breast cancer overdetection with mammography screening: a systematic review. Lancet Oncol. 2007;812:1129–1138. [DOI] [PubMed] [Google Scholar]
- 9.Etzioni R, Gulati R, Mallinger L et al. . Influence of study features and methods on overdiagnosis estimates in breast and prostate cancer screening. Ann Intern Med. 2013;15811:831–838. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Etzioni R, Gulati R. Oversimplifying overdiagnosis. J Gen Intern Med. 2014;299:1218–1220. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Puliti D, Duffy SW, Miccinesi G et al. . Overdiagnosis in mammographic screening for breast cancer in Europe: a literature review. J Med Screen. 2012;19(suppl 1):42–56. [DOI] [PubMed] [Google Scholar]
- 12.Feuer EJ, Wun LM. How much of the recent rise in breast cancer incidence can be explained by increases in mammography utilization? A dynamic population model approach. Am J Epidemiol. 1992;13612:1423–1436. [DOI] [PubMed] [Google Scholar]
- 13.Howlader N, Noone A, Krapcho M et al. eds. SEER Cancer Statistics Review, 1975–2011. Bethesda, MD: National Cancer Institute; 2014. [Google Scholar]
- 14.Sullivan PD, Christine B, Connelly R et al. . Analysis of trends in age-adjusted incidence rates for 10 major sites of cancer. Am J Public Health. 1972;628:1065–1071. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Duffy SW, Chen HH, Tabar L et al. . Estimation of mean sojourn time in breast cancer screening using a Markov chain model of both entry to and exit from the preclinical detectable phase. Stat Med. 1995;1414:1531–1543. [DOI] [PubMed] [Google Scholar]
- 16.Shen Y, Zelen M. Screening sensitivity and sojourn time from breast cancer early detection clinical trials: mammograms and physical examinations. J Clin Oncol. 2001;1915:3490–3499. [DOI] [PubMed] [Google Scholar]
- 17.Shen Y, Zelen M. Robust modeling in screening studies: estimation of sensitivity and preclinical sojourn time distribution. Biostatistics. 2005;64:604–614. [DOI] [PubMed] [Google Scholar]
- 18.Draisma G, Etzioni R, Tsodikov A et al. . Lead time and overdiagnosis in prostate-specific antigen screening: importance of methods and context. J Natl Cancer Inst. 2009;1016:374–383. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Hakama M, Auvinen A, Day NE et al. . Sensitivity in cancer screening. J Med Screen. 2007;144:174–177. [DOI] [PubMed] [Google Scholar]
- 20.de Gelder R, Heijnsdijk EA, van Ravesteyn NT et al. . Interpreting overdiagnosis estimates in population-based mammography screening. Epidemiol Rev. 2011;33:111–121. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Smith RA. Author's reply. J Am Coll Radiol. 2014;1111:1098–1099. [DOI] [PubMed] [Google Scholar]
- 22.Etzioni R, Gulati R. Recognizing the limitations of cancer overdiagnosis studies: a first step towards overcoming them. J Natl Cancer Inst. 2015;1083:pii: djv345. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Duffy SW, Parmar D. Overdiagnosis in breast cancer screening: the importance of length of observation period and lead time. Breast Cancer Res. 2013;153:R41. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Schröder FH, Roobol MJ. ERSPC and PLCO prostate cancer screening studies: what are the differences? Eur Urol. 2010;581:46–52. [DOI] [PubMed] [Google Scholar]
- 25.Schröder FH, Hugosson J, Roobol MJ et al. . Prostate-cancer mortality at 11 years of follow-up. N Engl J Med. 2012;36611:981–990. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Schröder FH, Hugosson J, Roobol MJ et al. . Screening and prostate cancer mortality: results of the European Randomised Study of Screening for Prostate Cancer (ERSPC) at 13 years of follow-up. Lancet. 2014;3849959:2027–2035. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Gulati R, Wever EM, Tsodikov A et al. . What if I don't treat my PSA-detected prostate cancer? Answers from three natural history models. Cancer Epidemiol Biomarkers Prev. 2011;205:740–750. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Kopans DB. Arguments against mammography screening continue to be based on faulty science. Oncologist. 2014;192:107–112. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Smith RA. Counterpoint: overdiagnosis in breast cancer screening. J Am Coll Radiol. 2014;117:648–652. [DOI] [PubMed] [Google Scholar]
- 30.Seigneurin A, Labarère J, Duffy SW et al. . Overdiagnosis associated with breast cancer screening: A simulation study to compare lead-time adjustment methods [published online ahead of print September 2, 2015] Cancer Epidemiol. (doi:10.1016/j.canep.2015.08.013doi:10.1016/j.canep.2015.08.013). [DOI] [PubMed] [Google Scholar]
- 31.Zelen M, Feinleib M. On the theory of screening for chronic diseases. Biometrika. 1969;563:601–614. [Google Scholar]
- 32.Pinsky PF. Estimation and prediction for cancer screening models using deconvolution and smoothing. Biometrics. 2001;572:389–395. [DOI] [PubMed] [Google Scholar]
- 33.Tsodikov A, Szabo A, Wegelin J. A population model of prostate cancer incidence. Stat Med. 2006;2516:2846–2866. [DOI] [PubMed] [Google Scholar]
- 34.Lansdorp-Vogelaar I, Gulati R, Mariotto AB et al. . Personalizing age of cancer screening cessation based on comorbid conditions: model estimates of harms and benefits. Ann Intern Med. 2014;1612:104–112. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Wu GH-M, Auvinen A, Yen AM-F et al. . The impact of interscreening interval and age on prostate cancer screening with prostate-specific antigen. Eur Urol. 2012;615:1011–1018. [DOI] [PubMed] [Google Scholar]
- 36.Zahl PH, Jørgensen KJ, Gøtzsche PC. Lead-time models should not be used to estimate overdiagnosis in cancer screening. J Gen Intern Med. 2014;299:1283–1286. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Etzioni R, Xia J, Hubbard R et al. . A reality check for overdiagnosis estimates associated with breast cancer screening. J Natl Cancer Inst. 2014;10612:pii: dju315. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Baker SG, Prorok PC, Kramer BS. Lead time and overdiagnosis. J Natl Cancer Inst. 2014;10612:pii: dju346. [DOI] [PubMed] [Google Scholar]
- 39.Otto SJ, Moss SM, Määttänen L et al. . PSA levels and cancer detection rate by centre in the European Randomized Study of Screening for Prostate Cancer. Eur J Cancer. 2010;4617:3053–3060. [DOI] [PubMed] [Google Scholar]