Skip to main content
British Journal of Cancer logoLink to British Journal of Cancer
. 2013 Dec 24;110(3):584–592. doi: 10.1038/bjc.2013.791

Comparison of cancer diagnostic intervals before and after implementation of NICE guidelines: analysis of data from the UK General Practice Research Database

R D Neal 1,*, N U Din 2, W Hamilton 3, O C Ukoumunne 4, B Carter 5, S Stapley 6, G Rubin 7
PMCID: PMC3915139  PMID: 24366304

Abstract

Background:

The primary aim was to use routine data to compare cancer diagnostic intervals before and after implementation of the 2005 NICE Referral Guidelines for Suspected Cancer. The secondary aim was to compare change in diagnostic intervals across different categories of presenting symptoms.

Methods:

Using data from the General Practice Research Database, we analysed patients with one of 15 cancers diagnosed in either 2001–2002 or 2007–2008. Putative symptom lists for each cancer were classified into whether or not they qualified for urgent referral under NICE guidelines. Diagnostic interval (duration from first presented symptom to date of diagnosis in primary care records) was compared between the two cohorts.

Results:

In total, 37 588 patients had a new diagnosis of cancer and of these 20 535 (54.6%) had a recorded symptom in the year prior to diagnosis and were included in the analysis. The overall mean diagnostic interval fell by 5.4 days (95% CI: 2.4–8.5; P<0.001) between 2001–2002 and 2007–2008. There was evidence of significant reductions for the following cancers: (mean, 95% confidence interval) kidney (20.4 days, −0.5 to 41.5; P=0.05), head and neck (21.2 days, 0.2–41.6; P=0.04), bladder (16.4 days, 6.6–26.5; P⩽0.001), colorectal (9.0 days, 3.2–14.8; P=0.002), oesophageal (13.1 days, 3.0–24.1; P=0.006) and pancreatic (12.6 days, 0.2–24.6; P=0.04). Patients who presented with NICE-qualifying symptoms had shorter diagnostic intervals than those who did not (all cancers in both cohorts). For the 2007–2008 cohort, the cancers with the shortest median diagnostic intervals were breast (26 days) and testicular (44 days); the highest were myeloma (156 days) and lung (112 days). The values for the 90th centiles of the distributions remain very high for some cancers. Tests of interaction provided little evidence of differences in change in mean diagnostic intervals between those who did and did not present with symptoms specifically cited in the NICE Guideline as requiring urgent referral.

Conclusion:

We suggest that the implementation of the 2005 NICE Guidelines may have contributed to this reduction in diagnostic intervals between 2001–2002 and 2007–2008. There remains considerable scope to achieve more timely cancer diagnosis, with the ultimate aim of improving cancer outcomes.

Keywords: diagnostic interval, NICE guidelines, urgent referral, earlier diagnosis, database, cancer, diagnosis, symptoms, primary care, cohort


Earlier diagnosis of cancer is increasingly acknowledged as a key element of the drive to improve cancer outcomes (Department of Health, 2011). An estimated 5–10 000 deaths within 5 years of diagnosis could be avoided annually in England if efforts to promote earlier diagnosis and appropriate primary surgical treatment are successful (Abdel-Rahman et al, 2009) and a National Awareness and Early Diagnosis Initiative (NAEDI) is addressing this challenge (Richards, 2009). Elsewhere in Europe, similar objectives are being pursued by a variety of national initiatives (Olesen et al, 2009).

One of the key phases in the journey for people with symptoms who go on to develop cancer is the ‘diagnostic interval'. This is the period between the first presentation of potential cancer symptoms (usually to primary care) and diagnosis (Hamilton, 2010; Weller et al, 2012). By contributing to shorter overall times to diagnosis, shorter diagnostic intervals should lead to earlier stage diagnoses and better cancer outcomes (Richards et al, 1999; Tørring et al, 2011), although the body of evidence to support this hypothesis remains limited (Neal, 2009). Expediting the assessment of patients with suspected cancer has been a priority for the UK Government since 1997 (Department of Health, 1997). An urgent referral pathway for suspected breast cancer was introduced in 1999 and for all other cancers in 2000. Following the publication in 2005 of NICE guidance on urgent referral for suspected cancer (National Institute for Health and Clinical Excellence, 2005), this pathway attained a new prominence because Primary Care Trusts were monitored by the Healthcare Commission on their implementation of NICE guidance. Urgent referral rates rose from approximately 350 out of 100 000 to 1900 out of 100 000 between 2004 and 2010 (Leeds North West PCT, 2005; National Cancer Information Network, 2011).The Cancer Reform Strategy (Department of Health, 2007) introduced a strong policy focus on earlier diagnosis of cancer and resulted in the NAEDI (http://www.cancerresearchuk.org/cancer-info/spotcancerearly/naedi/). This included work in general practice, including a national audit of cancer diagnosis in primary care in 2009/2010 (Rubin et al, 2011). A wide-ranging programme of engagement with primary care has developed, from improving consultation skills and decision support to practice cancer profiles (http://ncat.nhs.uk/our-work/diagnosing-cancer-earlier/gps-and-primary-care#). Other initiatives intended to shorten the diagnostic interval have included additional resources to improve access to diagnostic tests.

Measuring diagnostic intervals is important because it allows temporal and international comparisons, and may identify cancers where specific interventions to expedite diagnosis could be targeted, since diagnostic pathways vary greatly between cancers (Allgar and Neal, 2005). It can complement recent insights into the number of GP consultations prior to referral to provide a more complete picture of the diagnostic process (Lyratzopoulos et al, 2012). Measurement of diagnostic intervals is also necessary to determine the effect of those NAEDI interventions directed at the primary care part of the pathway to diagnosis.

Primary care data sets offer an important resource in the study of cancer diagnostic pathways. These data sets have previously been used to determine the positive predictive value of symptoms for cancer (Jones et al, 2007; Dommett et al, 2012) and to construct clinical decision support tools (Hamilton, 2009; Hippisley-Cox and Coupland, 2012). The primary aim of this study was to use routinely collected data to compare diagnostic intervals between two cancer cohorts, defined before and after the implementation of the 2005 NICE Referral Guidelines for Suspected Cancer. Secondary aims were to compare diagnostic intervals across cancers, and across different presenting symptoms.

Materials and methods

Data set

The General Practice Research Database (GPRD, but now the Clinical Practice Research Datalink – CPRD) is the world's largest computerised database of anonymised longitudinal medical records from primary care. These records include details of all consultations and diagnoses. We used the following patient inclusion criteria:

  • A new record of a primary diagnosis of 15 types of incident cancer (breast, lung, colorectal, gastric, oesophageal, pancreatic, kidney, bladder, testicular, cervical, endometrial, leukaemia, myeloma, lymphoma, and head and neck) in the study period. Three cancers (ovarian, brain, and prostatic) of the 18 adult cancers with the highest incidence were not studied because the data set was also being used to identify and quantify the risks of cancer with each cancer site, and these three sites had been previously studied for that purpose. Hence, diagnostic intervals for these three cancers were not studied.

  • At least 1 year of complete GPRD records before diagnosis.

  • Aged ⩾40 years at diagnosis. Younger patients were not included because of the rarity of cancer diagnoses in this group. This is in keeping with many similar primary care studies (e.g. Hamilton, 2009).

The first entry of a code pertaining to the cancer diagnosis was taken as the date of diagnosis and the clinical record for the 12-month period preceding this date was studied.

Patient cohorts

Two cohorts of patients were compared. Cohort 1 consisted of patients diagnosed between 01 January 2001 and 31 December 2002 inclusive, and Cohort 2 between 01 January 2007 and 31 December 2008 inclusive. These cohorts were chosen to allow sufficient time before and after the publication and implementation of the 2005 NICE Referral Guidelines for Suspected Cancer.

Symptom codes

A list of potential symptoms for each cancer was developed and agreed between RDN, GR, and WH, all practising clinicians. The principles adopted were as follows:

  • Symptoms were those of primary local and regional disease, not metastatic or recurrent disease.

  • Symptoms with a published independent association with cancer, and carrying a risk of >0.5% for a patient presenting to primary care, based upon:
    • systematic review evidence from primary care studies, or mixed primary/secondary care where primary care studies could be easily identified, or
    • single primary care studies using rigorous methods of the above, or
    • consensus statements.

These symptom lists were categorised as ‘site-specific symptoms'. In addition, ‘non-specific' symptoms, potentially caused by any cancer, were agreed, again with reference to the literature. These were: anaemia, anorexia, fatigue, and weight loss. These four symptoms were used in the analyses for all 15 cancers, along with their corresponding site specific symptoms. The full list of symptoms and further detail on applying codes to the data set and validation of identifying codes are shown in Appendix 1.

NICE-qualifying symptoms

We classified all symptoms according to whether they were ‘NICE-qualifying symptoms' or not. NICE-qualifying symptoms were those specifically cited in the NICE Guideline for Urgent Referral of Suspected Cancer as requiring urgent referral for either investigation or specialist assessment (National Institute for Health and Clinical Excellence, 2005). To do this, the three clinical researchers (GR, WH, and RDN) independently classified the list of symptoms for each cancer; these were compared and consensus reached. A number of assumptions had to be made in this process. These, along with the final lists are shown in Appendix 2.

Diagnostic intervals

The ‘diagnostic interval' was defined as the duration from the first occurrence of a symptom code in GPRD to the date of cancer diagnosis. The date of diagnosis was defined as the first entry of the code pertaining to a cancer diagnosis in the primary care record, in exactly the same way as many other studies of primary care diagnosis (e.g. Hamilton, 2009). We analysed data for 1 year before diagnosis. Although there have been reports of patients experiencing symptoms for more than a year before diagnosis (Corner et al, 2005), it is difficult to know whether the very early symptoms genuinely arise from the cancer, as many cancer symptoms may also arise from benign or incidental conditions. In the CAPER studies, no symptom was reliably more common in cases than controls more than a year before diagnosis in colorectal, lung or prostate cancers (Hamilton, 2009). Thus we chose 1 year as a reasonable compromise, to minimise the risk of our mislabelling a symptom actually unrelated to the cancer as being the index symptom. Our definitions and methods are in keeping with recently published recommendations (Weller et al, 2012).

Data analysis

Mean (s.d.) patient age and the percentage of females are reported for each cancer type within each cohort. For each cancer–cohort combination, the percentages of patients who had any identifiable symptom code during the year prior to diagnosis are presented. Diagnostic interval was calculated only for those patients who had identifiable symptom codes. For each cancer–cohort combination, the distribution of diagnostic interval was summarised for first symptomatic presentation, reporting the mean, standard deviation, median, inter-quartile range (IQR), and 90th centile. Median, IQR, and 90th centiles are shown as the preferred method for describing these skewed data but the t-test was used to compare the mean diagnostic intervals between Cohort 1 (2001–2002) and Cohort 2 (2007–2008), both overall and for each cancer type, as we wanted to make inferences about the mean change. Therefore, the mean and standard deviation for each cohort are also shown (Thompson and Barber, 2000). Because the diagnostic interval distributions were skewed, we validated the t-test results by constructing bias corrected accelerated bootstrap confidence intervals for the mean difference as these are robust to non-Normality (Davison and Hinkley, 1997). As the bootstrap confidence intervals were virtually the same as the t-test confidence intervals, we report results from the latter analysis since it also provides P-values. Linear regression was used to carry out tests of interaction to compare the mean change in diagnostic interval between presentations of a NICE-qualifying symptom alone or in combination (‘NICE') and presentations of a non-NICE-qualifying symptom (‘not NICE'). These regression models included as predictor variables cohort status, NICE category status, and the interaction between cohort status and NICE category status. The P-value for the interaction term was used to quantify evidence that the change in mean diagnostic interval differs between NICE and non-NICE categories. All data manipulation and analyses were performed using Stata software version 10.

Results

Demographic characteristics and proportions of patients with recorded symptoms

In total, 37 588 patients had a new diagnosis of cancer (Cohort 1 – 15 906 and Cohort 2 – 21 682), and of these 20 535 (54.6%) had a recorded symptom in the year prior to diagnosis and were included in the analysis (Cohort 1 – 8181 and Cohort 2 – 12 354). The age and gender of the patients as well as the percentage with symptoms for each cancer group are summarised separately for the two cohorts (2001–2002 and 2007–2008) in Table 1. The mean ages of patients in the two cohorts were similar for all cancers. Because the data set only contained patients aged 40 years or more, the mean ages in our cohorts of those cancers that also affect younger people are artefactually high. The proportion of cases that were male increased for all cancers over time except for lung and pancreatic. The proportion of patients with recorded symptoms for each of the cancers increased between 2001–2002 and 2007–2008.

Table 1. Demographic characteristics of patients in 15 cancer sites.

 
 
 
Age at diagnosis
 
Cancer site Cohort N Mean s.d. % female % with recordedsymptoms
Colorectal Cohort 1 (2001–2002) 3163 71.4 11.3 45.9 61.6
 
Cohort 2 (2007–2008)
4377
71.8
11.2
44.7
67.7
Lung Cohort 1 (2001–2002) 2963 71.0 10.4 38.7 61.3
 
Cohort 2 (2007–2008)
4384
71.8
10.4
43.6
65.0
Breast Cohort 1 (2001–2002) 1717 65.5 12.7 100.0 47.4
 
Cohort 2 (2007–2008)
1987
66.6
13.2
100.0
50.1
Pancreatic Cohort 1 (2001–2002) 623 71.6 11.3 51.5 68.9
 
Cohort 2 (2007–2008)
789
72.5
11.8
53.1
71.4
Oesophageal Cohort 1 (2001–2002) 992 71.6 11.2 37.5 67.7
 
Cohort 2 (2007–2008)
1236
71.0
11.0
30.9
75.2
Gastric Cohort 1 (2001–2002) 972 72.8 11.3 38.0 55.0
 
Cohort 2 (2007–2008)
1341
73.1
10.8
33.0
56.8
Endometrial Cohort 1 (2001–2002) 502 66.9 11.6 100.0 46.2
 
Cohort 2 (2007–2008)
864
66.7
10.9
100.0
59.6
Cervix Cohort 1 (2001–2002) 277 60.6 14.5 100.0 27.8
 
Cohort 2 (2007–2008)
263
59.9
14.7
100.0
35.4
Kidney Cohort 1 (2001–2002) 412 67.2 11.6 39.6 29.1
 
Cohort 2 (2007–2008)
966
68.7
11.6
37.6
35.0
Bladder Cohort 1 (2001–2002) 1090 72.6 10.5 25.0 61.0
 
Cohort 2 (2007–2008)
1486
73.2
10.5
25.4
68.5
Testicular Cohort 1 (2001–2002) 92 51.0 10.7 NA 37.0
 
Cohort 2 (2007–2008)
108
49.9
10.3
NA
63.9
Head and Neck Cohort 1 (2001–2002) 392 68.3 12.4 32.9 39.5
 
Cohort 2 (2007–2008)
394
66.8
12.2
23.6
54.8
Lymphoma Cohort 1 (2001–2002) 1092 67.5 12.2 47.6 23.4
 
Cohort 2 (2007–2008)
1455
68.6
12.0
47.8
31.5
Leukaemia Cohort 1 (2001–2002) 1037 70.3 11.6 44.1 17.2
 
Cohort 2 (2007–2008)
1252
70.2
12.1
40.0
19.4
Myeloma Cohort 1 (2001–2002) 582 72.0 10.5 47.4 38.0
 
Cohort 2 (2007–2008)
780
71.8
10.8
41.9
43.7
Pooled Cohort 1 (2001–2002) 15 906 69.9 11.8 50.1 51.4
  Cohort 2 (2007–2008) 21 682 70.5 11.7 48.8 57.0

Diagnostic intervals

First presentation of any cancer-related symptom

The diagnostic intervals in 2001–2002 and 2007–2008 for each cancer are summarised in Table 2. There was a reduction in mean diagnostic interval of 5.4 days (95% CI: 2.4–8.5; P<0.001) from 2001–2002 to 2007–2008 for first presentation of any cancer symptom. There was significant evidence at the 5% level of reductions for six cancers (mean reduction; 95% confidence interval): kidney (20.4 days; −0.5 to 41.5), head and neck (21.2 days; 0.2–41.6), bladder (16.4 days; 6.6–26.5), colorectal (9.0 days; 3.2–14.8), oesophageal (13.1 days; 3.0–24.1), and pancreatic (12.6 days; 0.2–24.6). Median diagnostic intervals were longer for all cancers, except leukaemia and myeloma, in Cohort 1 compared with Cohort 2. For the 2007–2008 cohort, the cancers with the shortest median diagnostic intervals were breast (26 days), testicular (44 days), and oesophageal (58 days); and those with the longest were myeloma (156 days), lung (112 days), and lymphoma (99 days). Similarly, the cancers with the shortest 90th centile diagnostic intervals were testicular (113 days), breast (203 days), and cervical (232 days); and those with the longest were myeloma (336 days), lung (325 days), and gastric (315 days). The 90th centile diagnostic intervals were 4–7 months for both cohorts for breast and testicular cancers and >9 months for both cohorts for all other cancers (except cervical Cohort 2).

Table 2. Analysis of diagnostic intervals between cohorts 2001–2002 and 2007–2008 in 15 cancer sites by first presentation of any cancer symptom.
      Diagnostic interval (days)
 
 
 
 
 
IQR
 
 
 
 
95% CI
 
Cancer site Cohort N Median 25th percentile 75th percentile 90th percentile Mean s.d. Mean differencea Lower Upper P-value
Colorectal Cohort 1 (2001–2002) 1947 100 42 201 298 129.6 103.1 −9.0 −14.8 −3.2 0.002
 
Cohort 2 (2007–2008)
2962
80
37
187
298
120.6
103.9
 
 
 
 
Lung Cohort 1 (2001–2002) 1816 114 48 238 318 144.8 110.7 2.4 −4.1 8.8 0.47
 
Cohort 2 (2007–2008)
2851
112
45
251
325
147.2
114.0
 
 
 
 
Breast Cohort 1 (2001–2002) 813 31 14 72 210 65.5 84.7 −3.5 −11.5 4.1 0.37
 
Cohort 2 (2007–2008)
996
26
15
61
203
62.0
84.8
 
 
 
 
Pancreatic Cohort 1 (2001–2002) 429 67 29 163 291 108.3 103.1 −12.6 −24.6 0.2 0.04
 
Cohort 2 (2007–2008)
563
56
25
142
250
95.7
94.1
 
 
 
 
Oesophageal Cohort 1 (2001–2002) 672 72 31 189 304 116.9 107.4 −13.1 −24.1 −3.0 0.006
 
Cohort 2 (2007–2008)
930
58
28
154
283
103.7
102.4
 
 
 
 
Gastric Cohort 1 (2001–2002) 535 101 43 220 314 135.5 108.4 −6.0 −18.7 5.5 0.33
 
Cohort 2 (2007–2008)
761
89
36
206
315
129.5
109.5
 
 
 
 
Endometrial Cohort 1 (2001–2002) 232 85 43 154 240 109.6 86.8 −8.3 −21.8 5.3 0.22
 
Cohort 2 (2007–2008)
515
70
36
140
238
101.3
86.0
 
 
 
 
Cervix Cohort 1 (2001–2002) 77 75 37 162 295 110.2 96.6 −13.7 −41.8 14.1 0.33
 
Cohort 2 (2007–2008)
93
66
30
113
232
96.5
90.5
 
 
 
 
Kidney Cohort 1 (2001–2002) 120 120 50 211 294 136.2 101.0 −20.4 −41.5 0.5 0.05
 
Cohort 2 (2007–2008)
338
80
39
169
280
115.8
98.0
 
 
 
 
Bladder Cohort 1 (2001–2002) 664 107 52 214 300 134.4 101.0 −16.4 −26.5 −6.6 <0.001
 
Cohort 2 (2007–2008)
1018
81
38
178
290
118.1
102.8
 
 
 
 
Testicular Cohort 1 (2001–2002) 34 53 25 104 190 75.6 68.7 −19.2 −45.1 5.7 0.14
 
Cohort 2 (2007–2008)
69
44
21
72
113
56.5
53.0
 
 
 
 
Head and Neck Cohort 1 (2001–2002) 155 113 59 212 322 141.7 107.1 −21.2 −41.6 −0.2 0.04
 
Cohort 2 (2007–2008)
216
87
50
177
272
120.4
92.9
 
 
 
 
Lymphoma Cohort 1 (2001–2002) 288 109 50 229 330 143.2 110.8 −11.1 −28.5 4.3 0.18
 
Cohort 2 (2007–2008)
458
99
44
218
302
132.1
104.7
 
 
 
 
Leukaemia Cohort 1 (2001–2002) 178 88 27 199 324 125.3 112.7 3.0 −20.0 25.6 0.80
 
Cohort 2 (2007–2008)
243
92
24
230
311
128.3
115.0
 
 
 
 
Myeloma Cohort 1 (2001–2002) 221 144 56 264 325 161.5 112.3 6.8 −12.4 26.4 0.50
 
Cohort 2 (2007–2008)
341
156
59
273
336
168.3
114.7
 
 
 
 
Pooled Cohort 1 (2001–2002) 8181 90 36 202 304 125.8 106.5 −5.4 −8.5 −2.4 <0.001
  Cohort 2 (2007–2008) 12 354 77 34 195 302 120.4 106.8        
a

‘−'sign denotes a reduction in diagnostic intervals from Cohort 2001–2002 to Cohort 2007–2008.

Differences by NICE-qualifying symptom category

For most of the cancers, patients in both cohorts who presented with NICE-qualifying symptoms had shorter diagnostic intervals than those who did not (gastric, cervical, and kidney cancer in Cohort 1 being the exceptions). Tests of interaction provided little evidence of differences between the NICE categories with respect to change in mean diagnostic interval between the two cohorts, with the exception of oesophageal cancer (P-value for interaction test=0.03), where there was a 16.8-day reduction for the NICE-qualifying symptom group and a 39.4-day increase for the non-NICE symptom group, and cervical cancer (P-value for interaction test=0.006), where there was a 55.4-day reduction for the NICE-qualifying symptom group and a 22.0-day increase for the non-NICE symptom group (Table 3).

Table 3. Analysis of diagnostic intervals between cohorts 2001–2002 and 2007–2008 in 15 cancer sites by NICE-category symptomsa.
        Diagnostic interval (days)
 
 
 
 
 
 
IQR
 
 
 
 
95% CI
 
Cancer site Symptom category Cohort N Median 25th percentile 75th percentile 90th percentile Mean s.d. Mean differenceb Lower Upper Test of interaction P-values
Colorectal NICE Cohort 1 (2001–2002) 1134 91 39 178 278 119.2 98.0 −9.5 −16.9 −2.0 0.81
    Cohort 2 (2007–2008) 1729 69 35 163 279 109.7 98.8        
  Not NICE Cohort 1 (2001–2002) 813 118 50 233 315 144.1 108.2 −8.1 −18.1 1.0  
 
 
Cohort 2 (2007–2008)
1233
104
43
222
312
136.0
109.0
 
 
 
 
Lung NICE Cohort 1 (2001–2002) 1651 109 47 232 318 142.3 110.6 3.5 −3.5 10.0 0.22
    Cohort 2 (2007–2008) 2549 109 43 251 326 145.8 114.4        
  Not NICE Cohort 1 (2001–2002) 165 171 70 267 314 169.6 109.0 −10.3 −31.5 10.6  
 
 
Cohort 2 (2007–2008)
302
142
63
257
323
159.3
109.6
 
 
 
 
Breast NICE Cohort 1 (2001–2002) 712 26 13 55 129 51.6 69.8 −9.8 −16.5 −3.6 0.74
    Cohort 2 (2007–2008) 822 22 14 42 87 41.9 58.8        
  Not NICE Cohort 1 (2001–2002) 101 148 69 257 322 163.5 112.1 −6.5 −34.5 21.6  
 
 
Cohort 2 (2007–2008)
174
137
48
261
342
157.0
118.1
 
 
 
 
Pancreatic NICE Cohort 1 (2001–2002) 397 64 28 155 291 106.0 103.3 −13.1 −26.5 −0.6 0.96
    Cohort 2 (2007–2008) 510 54 24 133 248 92.8 93.6        
  Not NICE Cohort 1 (2001–2002) 32 128 45 201 298 137.3 98.2 −14.2 −57.5 27.7  
 
 
Cohort 2 (2007–2008)
53
102
39
182
277
123.1
96.1
 
 
 
 
Oesophageal NICE Cohort 1 (2001–2002) 648 70 31 188 304 115.2 106.8 −16.8 −27.3 −6.7 0.03
    Cohort 2 (2007–2008) 882 55 27 141 269 98.4 99.2        
  Not NICE Cohort 1 (2001–2002) 24 154 58 275 354 161.7 115.1 39.4 −16.4 92.6  
 
 
Cohort 2 (2007–2008)
48
204
94
305
355
201.0
113.0
 
 
 
 
Gastric NICE Cohort 1 (2001–2002) 496 101 42 221 314 134.9 109.1 −9.4 −22.5 2.5 0.18
    Cohort 2 (2007–2008) 680 85 36 202 312 125.5 107.9        
  Not NICE Cohort 1 (2001–2002) 39 101 61 214 285 142.5 100.3 20.3 −22.8 58.5  
 
 
Cohort 2 (2007–2008)
81
135
54
266
342
162.8
117.5
 
 
 
 
Endometrial NICE Cohort 1 (2001–2002) 184 76 37 128 209 95.5 76.1 −11.1 −24.2 2.2 0.92
    Cohort 2 (2007–2008) 393 62 34 108 188 84.9 71.1        
  Not NICE Cohort 1 (2001–2002) 48 149 78 219 338 163.6 103.8 −9.2 −44.4 23.4  
 
 
Cohort 2 (2007–2008)
122
133
65
235
310
154.4
106.3
 
 
 
 
Cervical NICE Cohort 1 (2001–2002) 43 82 41 185 305 118.6 101.2 −55.4 −91.6 −22.4 0.006
    Cohort 2 (2007–2008) 40 40 28 87 146 63.2 55.7        
  Not NICE Cohort 1 (2001–2002) 34 68 33 162 224 99.6 90.8 22.0 −20.9 62.4  
 
 
Cohort 2 (2007–2008)
53
84
49
197
292
121.6
103.3
 
 
 
 
Kidney NICE Cohort 1 (2001–2002) 67 120 49 175 280 134.2 96.7 −30.0 −56.1 −3.6 0.33
    Cohort 2 (2007–2008) 179 71 37 144 263 104.1 93.1        
  Not NICE Cohort 1 (2001–2002) 53 119 50 216 295 138.7 107.0 −9.7 −41.7 23.1  
 
 
Cohort 2 (2007–2008)
159
98
42
201
290
129.0
101.8
 
 
 
 
Bladder NICE Cohort 1 (2001–2002) 588 98 48 199 284 128.1 99.3 −19.2 −28.7 −9.1 0.46
    Cohort 2 (2007–2008) 878 71 35 154 284 108.9 99.3        
  Not NICE Cohort 1 (2001–2002) 76 199 105 267 317 182.8 100.7 −7.9 −38.9 20.1  
 
 
Cohort 2 (2007–2008)
140
178
86
268
330
174.9
106.7
 
 
 
 
Testicular NICE Cohort 1 (2001–2002) 19 41 25 60 71 42.0 22.4 1.1 −11.3 14.7 0.09
    Cohort 2 (2007–2008) 45 41 21 57 76 43.1 30.0        
  Not NICE Cohort 1 (2001–2002) 15 168 38 190 203 118.3 83.6 −36.8 −81.2 15.6  
 
 
Cohort 2 (2007–2008)
24
51
25
112
196
81.5
74.7
 
 
 
 
Head and Neck NICE Cohort 1 (2001–2002) 137 110 58 211 322 139.1 106.8 −20.5 −44.1 1.9 0.77
    Cohort 2 (2007–2008) 186 86 50 170 271 118.7 91.7        
  Not NICE Cohort 1 (2001–2002 18 136 69 259 337 160.9 110.8 −29.5 −88.3 34.6  
 
 
Cohort 2 (2007–2008)
30
89
47
204
280
131.4
100.5
 
 
 
 
Pooled NICE Cohort 1 (2001–2002) 6747 84 34 193 300 120.8 105.5 −6.2 −9.6 −3.0 0.88
    Cohort 2 (2007–2008) 9906 71 31 182 295 114.6 105.2        
  Not NICE Cohort 1 (2001–2002 1434 128 54 238 314 149.5 107.8 −5.6 −12.8 1.1  
    Cohort 2 (2007–2008) 2448 115 48 235 318 143.9 109.7        
a

Data for lymphoma, leukaemia, and myeloma not shown because the numbers were too small to compare changes in diagnostic intervals between the NICE categories.

b

‘−'sign denotes a reduction in diagnostic intervals from Cohort 2001–2002 to Cohort 2007–2008.

Discussion

For a group of 15 cancers, time from first presentation to the general practitioner to diagnosis reduced between 2001–2002 and 2007–2008. The size of the reduction differed across cancers. The values for the 90th centiles of the distributions remain very high for some cancers, and indeed increased for four cancer types. for the 2007–2008 cohort, median diagnostic intervals remained >2 months for 10 of the 15 cancers studied, while for 13 of the 15 cancers, 1 in 10 patients had a diagnostic interval of over 7 months. It is reasonable to suggest that these findings have clinical significance for some of the cancers. There were large differences in diagnostic intervals among cancer sites. There is now good evidence, using robust methods, for better survival in colorectal cancer with shorter diagnostic intervals (Tørring et al, 2011). This is likely to be true in other cancers also. Hence, our view is that even modest reductions in diagnostic intervals (such as shown in our paper for some, but not all, of the cancers studied), across large populations, are likely to make a difference in stage and survival to some patients. This is clearly accepted at policy level (for example, for NAEDI in England) and it is estimated that about half of the difference in survival is due to ‘late diagnosis' (Abdel-Rahman et al, 2009). For the cancers where there was no or minimal change, an alternative explanation is that extensive efforts to improve diagnostic times over this time period were unsuccessful.

There are two bodies of evidence of relevance to this paper. These are previous studies of the duration of cancer diagnostic intervals, and the effects of interventions to reduce cancer diagnostic time. For the former we were aware of the paucity of evidence in this area, given the lack of past interest in the concept of the ‘diagnostic interval'. For the latter, there has been a recent systematic review that has addressed this topic (Mansell et al, 2011). This included 22 studies reporting interventions (predominantly educational) to reduce ‘primary care delay', with a variety of outcomes. Although some of these did report positive effects, for example on diagnostic accuracy, none of the included papers reported any measures of timeliness or delay.

Direct comparisons of diagnostic intervals with previous studies and with other countries are difficult for two reasons: first because of differences in the measurement and definition of ‘diagnostic intervals' (Weller et al, 2012), and second because of the dearth of the literature. We are aware of only one recent feasibility study that has reported diagnostic intervals per se (Murchie et al, 2012); this is because the diagnostic interval is a recent concept, but one that we think is important because it is modifiable and relatively easy to measure. Many other studies have reported other time intervals of the diagnostic journey but this is the first to report diagnostic intervals across different time periods on such a scale and in 15 cancers. A recent systematic review of interventions to reduce primary care delay reported no data on the duration of diagnostic intervals (Mansell et al, 2011).

Our results are in keeping with previous findings that suggest fast-track referrals may perversely lengthen waiting times for some patients routinely referred for suspected breast cancer (Potter et al, 2007) and may prioritise those with advanced disease in lung cancer, who are more likely to have ‘red flag' symptoms (Allgar et al, 2006). For oesophageal and gastric cancers, our findings may reflect changes in clinical practice and reduced use of gastroscopy resulting from the 2005 NICE guidance on dyspepsia management (National Institute for Health and Clinical Excellence, 2004). The 2005 Referral Guidelines for Suspected Cancer (National Institute for Health and Clinical Excellence, 2005) were a major revision of the initial Department of Health guidelines in 2000 (Department of Health, 2000), and were implemented widely in primary and secondary care. It is entirely plausible that, augmented by service redesign – in particular 2-week clinics, some of which were established before 2005, but which were fully established by 2005 – they have contributed to the falling diagnostic intervals. While we cannot draw conclusions about causality, we suggest that change may have resulted, at least in part, from implementation of the 2005 NICE guidelines. It is likely that an increased awareness of symptoms and symptom clusters in primary care has led to earlier referral for specialist opinion or diagnostic investigation, although more streamlined diagnostic processes in secondary care may also have had an influence.

The main strength of this study is that it uses a large, longitudinal, high-quality, and validated UK general practice data set, that has previously been used for cancer diagnostic studies (Jones et al, 2007; Dommett et al, 2012); and recent systematic reviews have confirmed the validity of diagnostic coding within GPRD (Herrett et al, 2010; Khan et al, 2010). While there are potential methodological issues in measuring diagnostic intervals, a recent consensus statement (Weller et al, 2012) makes recommendations on the design of studies recording the first presentation of symptoms, the use of primary care databases being recommended. Our definitions and reporting are in keeping with these recommendations. Our findings regarding the numbers of patients with recorded symptoms are compatible with the proportion of patients diagnosed as emergency presentations (Elliss-Brookes et al, 2012). The study specifically relates to the diagnostic interval only – the time period when diagnostic activity takes place; hence it informs the development of interventions to reduce this.

There are a number of limitations to this study and the findings must be interpreted with some caution. First the study design does not permit us to infer causality, and only reports an association (although a very plausible one). A number of changes in policy and practice may have contributed to changes in diagnostic intervals over time, the implementation of the 2005 NICE Referral Guidelines for Suspected Cancer being only one of them. Secondly, this study was dependent upon coded symptoms, and it is inevitable that some symptoms were not recorded, or were recorded in an inaccessible field (so-called ‘free-text'). Recent GPRD studies, however, indicate that free-text data usually just confirms that which has been entered in a coded (and therefore accessible) form (Tate et al, 2011), and electronic records have been found to be similar to paper records (Hamilton et al, 2003). Furthermore, there will be some cancers that presented with symptoms that had not been included in our defined list; these patients would not have been included in our analysis and are likely to have different patterns of presentation. We have also assumed that the symptom identified in the record was caused by the cancer, when it may have been co-incidental. Although we were not able to specifically identify screen-detected patients, it is likely that they would have had no symptoms, and would therefore have been excluded. Third, we chose to apply a cut-off point for symptoms of 12 months prior to the date of diagnosis. Some diagnostic intervals may have been longer than this, although the likely effect is small. Had we prolonged the duration we would have captured both more patients with genuine diagnostic intervals of greater than 1 year, and more patients with symptoms that were unrelated to their subsequent cancer diagnosis. This is also an area where there may be variation between cancers; however, for consistency and because there are no methodological precedents we used the time period of 12 months for all of the cancers. Fourth, caution needs to be applied to the interpretation of data for some specific cancer sites or groups. For example, people under the age of 40 years were not included in the data sets. This was a practical decision because most cancers are rare below this age and, when they do occur, may be atypical, for example being part of a familial syndrome. Heterogeneity within certain cancer groups (leukaemia, head and neck), may also limit the generalisability of our findings. All of these issues, however, are mitigated by the fact that they would have affected both cohorts in similar, but not identical ways, since the numbers of patients and the numbers of symptoms in the cohorts vary. Last, GPRD did not (at the time of analysis) permit linkage with hospital data.

We have found that diagnostic intervals for cancer in England reduced between 2001–2002 and 2007–2008. We propose that the implementation of cancer referral guidelines may have had some influence. Within each cohort, the contrast between the diagnostic intervals for patients with and without non-NICE-qualifying symptoms is stark. The findings do not tell us, however, that NICE-qualifying symptoms are necessarily the right symptoms to prioritise. For example, we know that mild anaemia as a first symptom of colorectal cancer has a higher mortality than severe anaemia (Stapley et al, 2006), the implication being that ‘softer' symptoms may go undetected for longer. Indeed a fast-track system may disadvantage patients who do not fulfil the criteria (Jones et al, 2001; Allgar et al, 2006). There is still a considerable challenge for policy and practice to further expedite diagnosis. The median diagnostic intervals vary considerably between cancers, and some remain very long. The 90th centile of the distribution remains very long in some cancers. These findings, when compared with data from studies of other measures of time intervals in the diagnostic process (Allgar and Neal, 2005; Hansen et al, 2011), show that delays in diagnosis happen at different stages in different cancers, hence interventions to expedite diagnosis need to be carefully tailored. Expediting diagnoses for ‘red flag' symptoms may be relatively straightforward; expediting it for all symptoms remains a challenge. At present, a range of initiatives is being implemented in general practice in order to achieve this (Rubin et al, 2011). We have described a method of measuring the diagnostic interval and have provided a baseline against which these initiatives can be assessed.

Acknowledgments

This research was funded by the National Cancer Action Team and the Department of Health Cancer Policy Team. The views contained in it are those of the authors and do not represent Department of Health policy. We can confirm that the corresponding author has had full access to the data and final responsibility for the decision to submit for publication. We would like to thank Rosemary Tate for early input into the protocol, staff of the GPRD for help in understanding the data. OCU is supported by the Peninsula Collaboration for Leadership in Applied Health Research and Care. Ethical approval: Independent Scientific Advisory Committee, numbers 09_0110 and 09_0111.

Author Contributions

The study was designed by RDN, GR and WH. NUD conducted the analysis with input from BC and OCU. SS helped in the preparation of the data set for analysis. RDN wrote the first draft of the paper. All authors commented on the manuscript. RDN will act as guarantor.

Appendix 1

List of cancer specific symptoms, by cancer; and process of applying Read codes to cancer site specific and non-specific symptoms

graphic file with name bjc2013791i1.jpg

The process of applying Read codes to cancer site specific and non-specific symptoms was developed using the Medical Dictionary Version 1.3.0 provided by GPRD. Different combinations of Read terms using both free text or specific medical term and wild card truncation, were used to maximise the scope of terms, and a list of possible terms prepared. The list was screened independently by GR, WH, and RDN and agreed. Symptom codes that had specific non-cancer causes mentioned in the code description were excluded unless they had a question mark. For example, ‘A083.11Diarrhoea & vomiting -? Infect' was included, whereas ‘A082.11Travellers' diarrhoea' was excluded. These lists were then used to identify the symptom codes for the period of 1 year prior to diagnosis (the lists are available from the authors on request). Two researchers (ND, SS) independently applied this process of identifying the symptom codes from the data set for a pilot cancer site (pancreatic) to compare and cross-validate the process of capturing the symptom codes. There was an 80.2% agreement between the two (Kappa statistics: score 0.62, 95% CI 0.51–0.73, P<0.001). The disagreement was on codes that were very vague/broad and almost invariably had very few occurrences. We have considerable experience in creating libraries of codes for symptoms, particularly in the field of cancer.

Appendix 2

NICE qualifying symptoms and assumptions made in their classification

graphic file with name bjc2013791i2.jpg

A number of assumptions had to be made in this process. These were as follows:

  • Symptoms with a qualifying symptom duration. For example, several symptoms qualify if they are ‘persistent' or present for a specific number of weeks. In these circumstances, we had to make the assumption that the symptom always fulfilled the duration.

  • Symptoms with a qualifying age. For example, dyspepsia for upper gastrointestinal cancers in patients aged over 55. In these circumstances we applied this age criterion to the data set.

  • Symptoms with a qualifying description. For example, breast lumps are qualifying if they are, ‘hard', ‘tethered', or ‘present after next period'. They are, however, rarely coded in such detail, hence we included ‘breast lump' (and its synonyms) as a qualifying symptom, without further description.

  • Symptoms needing qualification by another. There were two instances of this. ‘Unexplained upper abdominal pain with weight loss' for pancreatic cancer, and ‘Persistent urinary tract infections with haematuria'. We agreed that abdominal pain and urinary tract infections, alone, would be qualifying symptoms for their respective cancers.

  • Implicit qualifying symptom. ‘Anaemia' is not strictly listed as a qualifying symptom for haematological cancers, even though there is much reference to it in the guideline. We agreed that it should be classified as such.

References

  1. Abdel-Rahman M, Stockton D, Rachet B, Hakulinen T, Coleman MP. What if cancer survival in Britain were the same as in Europe: how many deaths are avoidable. Br J Cancer. 2009;101:S115–S124. doi: 10.1038/sj.bjc.6605401. [DOI] [PMC free article] [PubMed] [Google Scholar]
  2. Allgar VL, Neal RD. Delays in the diagnosis of six cancers: analysis of data from the National Survey of NHS Patients: Cancer. Br J Cancer. 2005;92:1959–1970. doi: 10.1038/sj.bjc.6602587. [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Allgar V, Neal RD, Ali N, Leese B, Heywood P, Proctor G, Evans J. Urgent general practitioner referrals for suspected lung, colorectal, prostate and ovarian cancer. Br J Gen Pract. 2006;56:355–362. [PMC free article] [PubMed] [Google Scholar]
  4. Corner J, Hopkinson J, Fitzsimmons D, Barclay S, Muers M. Is late diagnosis of lung cancer inevitable? Interview study of patients' recollections of symptoms before diagnosis. Thorax. 2005;60:314–319. doi: 10.1136/thx.2004.029264. [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Davison AC, Hinkley DV. Bootstrap Methods and their Application. Cambridge University Press: New York, USA; 1997. [Google Scholar]
  6. Department of Health . Cancer Reform Strategy. Department of Health: London; 2007. [Google Scholar]
  7. Department of Health . Referral Guidelines for Suspected Cancer. Department of Health: London; 2000. [Google Scholar]
  8. Department of Health . The New NHS: Modern, Dependable. Department of Health: London; 1997. [Google Scholar]
  9. Department of Health . Improving Outcomes: A Strategy For Cancer. Department of Health: London; 2011. [Google Scholar]
  10. Dommett RM, Redaniel MT, Stevens MCG, Hamilton W, Martin RM. Features of childhood cancer in primary care: a population-based nested case–control study. Br J Cancer. 2012;106:982–987. doi: 10.1038/bjc.2011.600. [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Elliss-Brookes L, McPhail S, Greenslade M, Shelton J, Hiom S, Richards M. Routes to diagnosis for cancer – determining the patient journey using multiple routine data sets. Br J Cancer. 2012;107:1220–1226. doi: 10.1038/bjc.2012.408. [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Hamilton WT, Round AP, Sharp D, Peters TJ. The quality of record keeping in primary care: a comparison of computerised, paper and hybrid systems. Br J Gen Pract. 2003;53:929–933. [PMC free article] [PubMed] [Google Scholar]
  13. Hamilton W. The CAPER studies: five case-control studies aimed at identifying and quantifying the risk of cancer in symptomatic primary care patients. Br J Cancer. 2009;101:S80–S86. doi: 10.1038/sj.bjc.6605396. [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Hamilton W. Cancer diagnosis in primary care. Br J Gen Pract. 2010;60:121–128. doi: 10.3399/bjgp10X483175. [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Hansen RP, Vedsted P, Sokolowski I, Sondergaard J, Olesen F. Time intervals from first symptom to treatment of cancer. A cohort study of 2212 newly diagnosed cancers patients. BMC Health Services Res. 2011;11:284. doi: 10.1186/1472-6963-11-284. [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Herrett E, Thomas SL, Schoonen WM, Smeeth L, Hall AJ. Validation and validity of diagnoses in the General Practice Research Database: a systematic review. Br J Clin Pharmacol. 2010;69:4–14. doi: 10.1111/j.1365-2125.2009.03537.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Hippisley-Cox J, Coupland C. Identifying women with suspected ovarian cancer in primary care: derivation and validation of algorithm. BMJ. 2012;344:d8009. doi: 10.1136/bmj.d8009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Jones R, Rubin G, Hungin P. Is the two week rule for cancer referrals working. BMJ. 2001;322:1555–1556. doi: 10.1136/bmj.322.7302.1555. [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Jones R, Latinovic R, Charlton J, Gulliford MC. Alarm symptoms in early diagnosis of cancer in primary care: cohort study using the General Practice Research Database. BMJ. 2007;334:1040. doi: 10.1136/bmj.39171.637106.AE. [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Khan NF, Harrison SE, Rose PW. Validity of diagnostic coding within the General Practice Research Database: a systematic review. Br J Gen Pract. 2010;60:e128–e136. doi: 10.3399/bjgp10X483562. [DOI] [PMC free article] [PubMed] [Google Scholar]
  21. Leeds North West PCT 2005Cancer Equity Audit Leeds North West PCTpp 12-13. Available at: http://www.york.ac.uk/yhpho/documents/hea/Website/NW%20PCT%20Cancer%20Equity%20Audit_Copy.pdf (accessed on 10/12/2013).
  22. Lyratzopoulos G, Neal RD, Barbiere JM, Rubin G, Abel G. Variation in the number of general practitioner consultations before hospital referral for cancer: findings from a national patient experience survey. Lancet Oncol. 2012;13:353–365. doi: 10.1016/S1470-2045(12)70041-4. [DOI] [PubMed] [Google Scholar]
  23. Mansell G, Shapley M, Jordan JL, Jordan K. Interventions to reduce primary care delay in cancer referral: a systematic review. Br J Gen Pract. 2011;61:e821–e835. doi: 10.3399/bjgp11X613160. [DOI] [PMC free article] [PubMed] [Google Scholar]
  24. Murchie P, Campbell NC, Delaney EK, Dinant G-J, Hannaford PC, Johansson L, Lee AJ, Rollano P, Spigt M. Comparing diagnostic delay in cancer: a cross-sectional study in three European countries with primary care-led health care systems. Family Pract. 2012;29:69–78. doi: 10.1093/fampra/cmr044. [DOI] [PubMed] [Google Scholar]
  25. National Cancer Information Network Urgent GP referral rates for suspected cancer. NCIN Data Briefing. 2011.
  26. National Institute for Health and Clinical Excellence . Dyspepsia. Management of dyspepsia in adults in primary care: London; 2004. [Google Scholar]
  27. National Institute for Health and Clinical Excellence . Referral Guidelines for Suspected Cancer. NICE: London; 2005. [Google Scholar]
  28. National Institute for Health and Clinical Excellence . The recognition and initial management of ovarian cancer. NICE: London; 2011. [PubMed] [Google Scholar]
  29. Neal RD. Do diagnostic delays in cancer matter. Br J Cancer. 2009;101:S9–12. doi: 10.1038/sj.bjc.6605384. [DOI] [PMC free article] [PubMed] [Google Scholar]
  30. Olesen F, Hansen RP, Vedsted P. Delay in diagnosis: the experience in Denmark. Br J Cancer. 2009;101:S5–S8. doi: 10.1038/sj.bjc.6605383. [DOI] [PMC free article] [PubMed] [Google Scholar]
  31. Potter S, Govindarajulu S, Shere M, Braddon F, Curran G, Greenwood R, Sahu AK, Cawthorn SJ. Referral patterns, cancer diagnoses, and waiting times after introduction of two week wait rule for breast cancer: prospective cohort study. BMJ. 2007;335:288–291. doi: 10.1136/bmj.39258.688553.55. [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Richards MA, Westcombe AM, Love SB, Littlejohns P, Ramirez AJ. Influence of delay on survival of patients with breast cancer: systematic review. Lancet. 1999;353:1119–1126. doi: 10.1016/s0140-6736(99)02143-1. [DOI] [PubMed] [Google Scholar]
  33. Richards MA. The National Awareness and Early Diagnosis Initiative in England: assembling the evidence. Br J Cancer. 2009;101:S1–S4. doi: 10.1038/sj.bjc.6605382. [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Rubin GP, McPhail S, Elliott K. National Audit of Cancer Diagnosis in Primary Care. Royal College of General Practitioners: London; 2011. [Google Scholar]
  35. Rubin G, Vedsted P, Emery J. Improving cancer outcomes: better access to diagnostics in primary care could be critical. Br J Gen Pract. 2011;61:317–318. doi: 10.3399/bjgp11X572283. [DOI] [PMC free article] [PubMed] [Google Scholar]
  36. Stapley S, Peters TJ, Sharp D, Hamilton W. The mortality of colorectal cancer in relation to the initial symptom and to the duration of symptoms: a cohort study in primary care. Br J Cancer. 2006;95:1321–1325. doi: 10.1038/sj.bjc.6603439. [DOI] [PMC free article] [PubMed] [Google Scholar]
  37. Tate AR, Martin AGR, Ali A, Cassell JA. Using free text information to explore how and when GPs code a diagnosis of ovarian cancer: an observational study using primary care records of patients with ovarian cancer. BMJ Open. 2011;1:e000025. doi: 10.1136/bmjopen-2010-000025. [DOI] [PMC free article] [PubMed] [Google Scholar]
  38. Thompson SG, Barber JA. How should costs data in pragmatic trials be analysed. BMJ. 2000;320:1197–1200. doi: 10.1136/bmj.320.7243.1197. [DOI] [PMC free article] [PubMed] [Google Scholar]
  39. Tørring ML, Frydenberg M, Hansen RP, Olesen F, Hamilton W, Vedsted P. Time to diagnosis and mortality in colorectal cancer: a cohort study in primary care. Br J Cancer. 2011;104:934–940. doi: 10.1038/bjc.2011.60. [DOI] [PMC free article] [PubMed] [Google Scholar]
  40. Weller D, Vedsted P, Rubin G, Walter F, Emery J, Scott S, Campbell C, Andersen RS, Hamilton W, Olesen F, Rose P, Nafees S, van Rijswijk E, Muth C, Beyer M, Neal RD. The Aarhus Statement: improving design and reporting of studies on early cancer diagnosis. Br J Cancer. 2012;106:1262–1267. doi: 10.1038/bjc.2012.68. [DOI] [PMC free article] [PubMed] [Google Scholar]

Articles from British Journal of Cancer are provided here courtesy of Cancer Research UK

RESOURCES