Abstract
Background
The lack of comparable metrics to evaluate prevention and early detection patient navigation programs impedes our ability to identify best practices.
Methods
The Prevention and Early Detection Workgroup of the Patient Navigation Leadership Summit was charged with making recommendations for common metrics specific to the prevention and early detection phase of the cancer care continuum. The workgroup began with a review of existing literature to characterize variability in published navigation metrics; then through discussion and group consensus developed a list of priority recommendations.
Results
Recommendations for researchers and program evaluators include:
Clearly document key program characteristics;
Use a set of core data elements to form the basis of your reported metrics; and
Prioritize data collection using methods with the least amount of bias
Conclusion
If navigation programs explicitly state the context of their evaluation and choose from among the core set of data elements, meaningful comparisons among existing programs should be feasible.
Keywords: patient navigation, prevention & control, early detection of cancer, cancer screening, program evaluation, measures
Introduction
Cancer control begins with primary and secondary prevention efforts which aim to reduce cancer incidence and advanced disease, respectively. The evidence is clear that certain cancers — those caused by tobacco use, viruses, or sun exposure, for example — can be prevented completely. Regular use of proven screening modalities, such as Pap tests for cervical cancer and colonoscopy for colorectal cancer, also result in prevention through the removal of precancerous lesions. Other screening tests can detect cancers of the breast, colon, rectum, cervix, prostate, oral cavity, and skin at early stages and translate into a direct mortality benefit when abnormal screening is followed by prompt diagnosis and treatment. Indeed, mounting evidence suggests that the delivery of prevention and early detection (PED) services are responsible for a substantial portion of the documented reduction in both cancer incidence1 and mortality1, 2 in the United States.
It is also well documented that not all populations benefit equally from these prevention efforts, in part because our current healthcare delivery system does not provide consistent, high-quality care to all.3 Whether defined by age, gender, race, insurance status, geographic location, or co-morbid medical condition, certain populations face significant barriers to accessing timely and quality cancer PED services consistently, if at all.4–6 Patient Navigation, which targets barriers faced by vulnerable populations in accessing timely, quality cancer care,7 was designed to address the critical disconnect between the discovery and delivery of life-saving cancer care services. In fact, the first patient navigation program was started in Harlem, New York in order to increase the delivery of mammography screening to Black women who were too often presenting with advanced cancer as a result of a lack of screening.8 This groundbreaking work used lay navigators from the local community to help at-risk women overcome barriers to accessing screening and diagnostic services and resulted in profound improvements in breast cancer care.8 Since then a growing number of studies documenting the promise of navigation have resulted in its widespread adoption as a means to deliver PED services.9–15
As navigation becomes integrated into standard cancer care services across the country,16 the lack of comparable metrics to evaluate these programs in different settings with diverse target populations impedes our ability to identify best practices and realize the full potential of this promising intervention. Thus, we aim here to provide recommendations for researchers and program evaluators to consider adopting when measuring the impact of their PED navigation programs. The intent is to facilitate consistent use of priority metrics, including process and intermediate outcome measure, that document the type and quality of work performed by prevention and early detection Patient Navigators (PN) working in diverse settings. Through the use of such measures, public health and health reform policies may be generated to provide reimbursement for services that ensure the delivery of timely, quality cancer prevention.
Methods
In March 2010, the American Cancer Society hosted the first National Patient Navigation Leadership Summit, where it convened cancer clinicians, researchers and practicing public health experts to develop a national evaluation agenda for patient navigation. The Prevention and Early Detection (PED) Work Group was charged with making recommendations for common metrics specific to the prevention and early detection phase of the cancer care continuum. The work group comprised 10 individuals, representing community-based organizations, clinical programs and academia with decades of experience implementing and evaluating patient navigation programs across diverse populations. The work group began with a review of existing literature to characterize variability in published metrics, then through discussion and group consensus developed a list of priority recommendations.
In early 2010 the Summit Planning Committee conducted a comprehensive review of the navigation literature to guide discussion at the March meeting. The PED work group updated the literature review in October of 2010. We searched the Pubmed database to identify original articles published any year, in English, using the key terms “patient navigation,” “patient navigator,” “navigation,” “navigator,” or “case management.” We also searched the references of each publication for additional relevant literature. In keeping with the scope of navigation as outlined by Dr. Freeman at the Summit, we included only intervention studies where navigators actively link patients to screening services. Educational or outreach navigation for the delivery of prevention education in community settings were excluded. We present here findings from the synthesis of thirty-two published articles we felt exemplified the breadth of published metrics. While not meant to be an exhaustive review of the extensive literature, the studies included are representative of the variability in existing metrics.
Findings from Literature Review
Most studies target breast and colorectal cancer, with fewer targeting cervical, lung and/or prostate cancer. Reported clinical outcome metrics fall into 2 discrete points along the continuum of cancer care: 1) Screening and 2) Diagnosis, while the remaining metrics focus on the processes specific to the navigation program. To date, no patient navigation intervention study has reported final endpoints such as survival or mortality. Rather, the current literature focuses mainly on intermediate clinical outcomes in the form of the delivery of recommended cancer prevention services. Only two studies17, 18 document a potential mortality benefit in the form of a stage shift at the time of diagnosis. As discussed below and summarized in Table 1, there exists wide variation in both the reporting of non-modifiable program characteristics as well as how study outcome metrics are defined and reported.
Table 1.
Variability in Prevention and Early Detection Metrics
Screening | |||||||||
---|---|---|---|---|---|---|---|---|---|
Study | Setting | Disease | Eligibility Criteria | Navigation mode |
Outcome Metric |
Process Metric |
Follow-Up Period |
Data Collection |
|
Burhansstipanov L, 2010 | Urban | Community | Breast | Age ≥ 39 Last mammo > 18 months |
Face-to-face Telephone |
Receipt of screening mammo | NR | NR | Self-report |
Burhansstipanov L, 2000 | Urban | Community | Breast | Age ≥39 Last mammo > 18 months |
Face-to-face Telephone |
Receipt of screening test | NR | NR | Self-report |
Ford, 2006 | Urban | Clinical | Prostate Lung Colorectal |
Age ≥55 | Telephone | Receipt of screening test (Next scheduled test) | NR | Time to next trial screening | Medical Record Review |
Dietrich, 2006 | Urban | Clinical | Breast Cervical Colorectal |
Age 50 – 69 Overdue at least 1 cancer screening |
Telephone | Adherence to recommended screening | NR | 18 months | Medical record review |
Fernandez, 2009 | Rural | Community | Breast Cervical |
Age ≥50 Farm worker status |
Face-to-face | Receipt of mammo, pap test | NR | 6 months | Self-report Medical record review for validity |
Wang, 2010 | Urban | Community | Cervical | Age 18+ Last pap > 12 months |
Face-to-face | Receipt of pap test | NR | 12 months | Self-report |
Dignan, 2005 | Urban | Community | Breast | Age 40+ Last mammo > 18 months |
Face-to-face Telephone |
Receipt of mammo | NR | 12 months | Self-report |
Han, 2009 | NR | Community | Breast | Age 40+ Last mammo > 24 months |
Face-to-face Telephone |
Receipt of mammo | NR | 6 months | Self-report |
Paskett, 2006 | Rural | Community | Breast | Age ≥ 41 Last mammo > 12 months |
Face-to-face follow-up telephone calls | Receipt of mammo | Barriers | 12 months | Medical record review |
Phillips, 2010 | Urban | Clinical | Breast | Age 51 – 70 Last mammo > 18 months |
Telephone | Adherence to recommended screening | NR | 24 months | Medical record review |
Weber, 1997 | Urban | Clinical | Breast | Age 52 – 77 Last mammo > 24 months |
Telephone | Receipt of mammo | NR | 10 months | Medical record review |
Clark, 2009 | Urban | Clinical | Breast | Age 18 – 75 | Telephone | Receipt of mammoMaintenance screening behavior Timely adherence to diagnostic resolution |
NR | 3 years | Medical record review |
Ma, 2009 | NR | Community | Colorectal | Age 50+ ACS guidelines for last CRC screening |
Face-to-face | Receipt of CRC screening | Barriers | 12 months | Self-report Verified with physician’s office |
Jandorf, 2005 | Urban | Clinical | Colorectal | Age 50+ FOBT > 1year FS or Barium Enema >5 years Colonoscopy >10 years |
Telephone | Receipt of CRC screening | NR | 6 months | Medical record review |
Lasser, 2009 | Urban | Clinical | Colorectal | Age 52 – 80 No CRC screening |
Telephone | Receipt of CRC screening | Number of contacts Hours of contact |
6 months | Medical record review |
Myers, 2008 | NR | Clinical | Colorectal | Age 50 – 79 Last practice visit < 24 months |
Telephone | Receipt of CRC screening | NR | 6 months | Self-report Medical record review |
Percac-Lima, 2008 | Urban | Clinical | Colorectal | Age 52 – 79 FOBT > 1year FS or Barium Enema >5 years Colonoscopy >10 years |
Telephone | Receipt of CRC screening | Barriers | 9 months | Medical record review |
Nash, 2006 | Urban | Clinical | Colorectal | All colonoscopies | NR | Receipt of colonoscopy | NR | 11 months | Medical record review |
Diagnosis | |||||||||
Study | Setting | Disease | Eligibility Criteria | Navigation mode | Outcome Metric | Process Metric | Follow-Up Period | Data Collection | |
Oluwole, 2003 | Urban | Clinical | Breast | Age: None Clinic patients |
NR | Stage at diagnosis | NR | Retrospective, cross-sectional | Medical record review |
Gabram, 2008 | Urban | Clinical | Breast | Age: None Clinic patients |
NR | Stage at diagnosis | NR | Retrospective, cross-sectional | Medical record review |
Ferrante, 2007 | Urban | Clinical | Breast | Age: ≥ 21 BIRADS 4 & 5 |
Telephone Face-to-face |
Timely adherence to diagnostic resolution Time to diagnostic resolution |
NR | 60 day interval Follow-up period unclear |
Medical record review |
Palmieri, 2009 | Urban | Clinical | Breast | Age: None 200% FPL |
NR | Timely adherence to diagnostic resolution Receipt of diagnostic resolution | NR | 60 day interval NR |
Medical record review |
Ell, Cancer Pract, 2002 | Urban | Clinical | Breast | Age: None Prescribed follow-up screening/diagnostic test |
Telephone | Receipt of diagnostic resolution Timely adherence to diagnostic resolution |
Barriers | NR ACR 4& 5: 60 day interval ACR 3: 240 day interval |
Appointment records Self-report |
Ell, 2007 | Urban | Hospital | Breast | Age: None ACR 3-5 |
Telephone | Receipt of diagnostic resolution Timely adherence to diagnostic resolution | Barriers | 8 months ACR 4&5: 60 days ACR 3: 240 days |
Medical record review |
Ell, J Women’s Health Gend Based Med 2002 | Urban | Clinical | Cervical | Age: None LGSIL & HGSIL Prescribed follow-up screening/ diagnostic test |
Telephone | Receipt of diagnostic resolution Timely adherence to diagnostic resolution |
NR | 12 months 30 day interval |
Medical record review |
Battaglia, 2007 | Urban | Clinical | Breast | Age: >18 Referred for evaluation |
Telephone | Timely adherence to diagnostic resolution | NR | ys | Medical record review Self-report |
Bastani, 2010 | Urban | Clinical | Breast | Age: None Referred to Surgery or Radiology for breast abnormality |
Telephone | Timely adherence to diagnostic resolution | NR | 6 months | Medical record review |
NR = Not recorded
Screening
We reviewed twenty navigation studies that targeted cancer screening as an outcome. We include studies with community-and clinically-based navigators in urban19–25 and rural settings.14, 26. The studies targeted diverse populations, including American Indians,14, 15, 19, 26 Korean-Americans,27 Chinese women,28 Latinas,14, 29 Blacks,14, 26, 30, 31 non-English speaking,32 poor Whites,14 low-income,32–34 Few programs were comprehensive targeting multiple cancer sites,29, 30, 33 while most target only one disease-specific screening.14, 15, 19–28, 31, 32, 34–36 Even among studies targeting the same disease, eligibility criteria for inclusion in programs vary, including the age of participants and the time since their last screening. For example, one mammography screening navigation study included women 52–77 years who had not had a mammogram in the previous 2 years,21 while another included women over 40 years whose last mammogram was only 12 or more months prior.26
Most studies document receipt of a screening test as the goal of navigation and report the outcome simply as a screening rate, defined as the proportion of eligible subjects who complete a recommended test, such as a mammogram, Pap test or colonoscopy, during the intervention period. The range of the intervention period across studies was wide, such that the time subjects were followed to assess the outcome varied from 6 months21, 24, 32–34, to 3 years.30 The most common follow-up period was six months.22, 23, 27, 29, 36 Two studies document adherence to recommended screening20, 33 as the goal of navigation and report the outcome as an adherence rate, defined as the proportion of eligible subjects who are up-to-date with a screening test as recommended by an existing guideline or standard. These two breast navigation studies differed in how they defined “adherent”; one utilized USPSTF guidelines,33 another HEDIS criteria.20 Only one study reports maintenance screening behavior31 which was defined as the percentage of annual mammograms that were actually obtained during the study period. Data collection methods were either self-reported behaviors13–15, 19, 27–29, 35 or objective evidence from medical record review.20–24, 26, 31–33, 36
Diagnosis
Of the thirteen studies included targeting the diagnostic phase of the cancer care continuum, 12 targeted breast cancer diagnosis8–10, 12, 13, 17, 18, 31, 37–40 while only one targeted cervical cancer.11 As above, we include studies with a range of program settings that target diverse populations. And similar to screening navigation studies, the range of the intervention period across studies was wide, and the time subjects were followed to assess the outcome varied. Studies report four clinical metrics at the point of diagnostic evaluation: (1) Receipt of diagnostic resolution,8–11, 38, 40 (2) time to diagnostic resolution,8, 37–39 (3) timely adherence to diagnostic resolution,9–13, 31, 37, 38 and less commonly (4) stage at diagnosis.17, 18
Five studies report receipt of diagnostic resolution8–11, 38, 40 as the goal of navigation. These studies present this outcome simply as a resolution rate, defined as the proportion of eligible subjects who complete diagnostic testing during the intervention period. The majority of studies reviewed report timeliness of diagnostic care as the goal of patient navigation. These studies documented time in two distinct ways: either (1) the time to diagnostic resolution8, 37–39 as a continuous outcome or (2) the rate of timely adherence to diagnostic resolution as a dichotomous outcome.9–13, 31, 37, 38
The most striking finding in reviewing these metrics is the lack of a consistent definition for what constitutes ‘diagnostic resolution’ or the diagnostic interval. Most studies use the date the abnormal screening test was performed as the index event or start date.9, 10, 31, 37, 38 However, there is widespread variability in the data point indicating diagnosis, diagnostic resolution or adherence to recommended follow up, ranging from the date of arrival to first diagnostic clinical visit12 to the actual date a tissue sample was obtained.8 When tissue diagnosis is not recommended, studies vary in reporting how a ‘diagnostic resolution’ is determined. For example, one study reports the endpoint as “until negative mammogram, benign biopsy, 6 month follow-up test, or start of cancer treatment”10 while other studies only include benign or malignant tissue as a diagnosis.39 There is similar variability in how investigators define “timely” ranging from 60 to 180 days.9, 13
The Patient Navigation Research Program, a collaborative multi-site research program designed to evaluate the efficacy of navigation after abnormal cancer screening, developed a set of “common” data points using the National Comprehensive Cancer Network (NCCN) guidelines as the major focus of clinical outcomes.41 While the results of this program are not yet published, the PNRP is the largest study to date on PED navigation. In their program, diagnostic resolution is defined as completion of the diagnostic test that results in a diagnosis or clinical evaluation that determines that no further evaluation is indicated. For example, a colonoscopy with biopsy confirming a malignant polyp or a colonoscopy in which no malignant lesion is identified would both serve as a diagnostic resolution.
The two studies reporting breast cancer stage at diagnosis as the outcome similarly reported population level data to assess the impact of a navigation program targeting individuals. 17, 18 While this suggests a potential impact of navigation, the methods used preclude a causative association to be determined.
Process metrics
In addition to intermediate clinical outcomes six studies studies included here report metrics that evaluate whether the intervention was implemented as intended. Five studies report navigator-documented barriers to care9, 10, 25, 31, 35. One study, by Lasser et al, documents the median number of contacts per patient and mean hours of telephone outreach per patient33. A descriptive study by Lin et al documented the types of barriers to care and time spent by the navigator.42 The PNRP is collecting the following process metrics in their multi-site program: barriers to care identified by navigator, actions taken by the navigator, and details of navigation encounters such as type of encounter and time spent.41 Only one33 of these studies have examined the association between these process measures with their outcomes, which represents an area in critical need for further study.
Recommendations for PED Metrics
In keeping with the goal of having a core set of priority metrics for navigation programs to measure impact on individuals and populations, it would be ideal to have consistent study characteristics, including eligibility criteria, follow up time intervals and outcome metrics. Obviously, this is not possible for several reasons. First, there are certain program characteristics that are inherently non-modifiable such as program setting and the populations they serve. In addition, the specific needs of populations appropriately dictate the intended outcomes of navigation, the ideal mode of navigator contact or specific navigator activities. Finally, there is wide variability in the amount and type of resources available for evaluation efforts. Community programs wishing to conduct an evaluation of their program may well have fewer resources than an externally funded research project such as the PNRP. Regardless, the existing literature illuminates the need for one approach that may bring us closer to this ideal. That is, the need for consistency in reporting both modifiable and non-modifiable program characteristics. Stating these clearly will facilitate meaningful program comparisons even in the absence of common outcome metrics.
Therefore, our first recommendation is to clearly document the following set of program characteristics:
1) Program setting
At a minimum knowledge of geographic settings such as urban, rural, suburban is an important distinction. More importantly, the system setting is essential to know when considering replicating a program. Beyond describing whether a program is community versus clinically-based, some detail on the specific area within a clinical or community setting is important, such as in primary care versus radiology.
2) Eligibility criteria of navigated subjects
These programmatic elements are necessary to order to interpret the outcomes and their potential impact for other populations. Most important are age, race/ethnicity, primary spoken language and time since last screening.
3) Mode of navigation
The primary mode of delivering the navigation program is a minimal program element essential to comparing study findings. Specifically, did the navigator interact with their target community in person (in a community setting, in a clinical setting) or on the telephone?
4) Time interval of the follow up period at which outcomes are assessed
This detail is critical to interpreting the meaning of a defined outcome. For example, it would important to know that two programs reporting a similar clinical outcome (ie: 90% of program participants completed their mammogram) have each measured their outcome at different time intervals (ie. one year versus 6 months).
Many of the observed differences in published PED outcomes are not in the data elements collected, rather in either the nomenclature used to describe them or the analyses used to report them. Thus, defining a core set of data elements, rather than firm outcome metrics is a much more realistic approach and comprises our second set of recommendations. Prioritizing the collection of these data elements will allow for the variability inherent in navigation programs that target different communities and systems of care while also allowing for meaningful comparisons. From these data elements, common metrics to represent prevention and early detection constructs can be created (see Tables 2, 3 and 4).
Table 2.
Recommended common data elements for Screening Metrics
Construct | Common Metrics | Core Data Elements | Analysis Strategies |
---|---|---|---|
Receipt of screening test | Completion of screening test (Yes/No) Timely completion of screening (Yes/No) Time to complete screening (# days) Kept appointment rates |
A. Date enrolled into navigation B. Date referred for screening C. Date test scheduled D. Date test completed E. Date test results are read/reported F. Date patient informed of test result Scheduled appointment status for each appointment: arrive, no show, reschedule, cancel |
Define “completion” i.e., “Yes” = a date is entered for D. Date test completed Define “time to complete screening” ie: Identify time intervals: # of days from A → B → C → D → E → F, A→ F, or any combination Define “timely” (i.e. NCCN guidelines) Define “kept appointment” & “not kept appointment” i.e. cancel vs reschedule vs no show |
Adherence to single recommended screening | Did not keep appointment rates Number of re-scheduled appointments |
C. Date most recent screening test completed |
Track for each scheduled appointment Analyzed at patient, provider, or practice level |
Adherence to recommended screening longitudinally | Adherent to routine recommended screening (Yes/No) | A. Name of professional guidelines that defines recommended screening maintenance (ie: USPSTF, NCCN) B. Date current screening test completed C. Date most recent screening test completed D. Date past screening tests completed |
Define “adherent” i.e. “Yes” = completed a series of tests over specified time period Clearly document time of follow up period |
•USFSTF= United States Preventive Services Task Force, NCCN= National Comprehensive Cancer Network
•HEDIS= Healthcare Effectiveness Data and Information Set
Table 3.
Recommended common data elements for Diagnostic Metrics
Construct | Common Metrics | Core Data Elements | Analysis Strategies |
---|---|---|---|
Receipt of diagnosis or resolution of screening abnormality | Completion of diagnostic resolution (Yes/No) Time to completion of diagnostic resolution42 (# days) Timely completion of diagnostic resolution (Yes/No) Kept appointment rates |
A. Date index screening test performed B. Date patient informed of test result C. Date enrolled into navigation clinical evaluation D. Date of first scheduled diagnostic test/clinic visit E. Date of completion of first diagnostic test/clinic visit F. Date of final scheduled diagnostic test/clinic visit G. Date of diagnostic resolution = date final test peformed H. Date diagnostic test read/reported I. Date patient informed of test results Scheduled appointment status for each appointment: arrive, no show, reschedule, cancel |
Define “diagnostic resolution” ie.“Yes”= a date exist for the F. “date of diagnostic resolution” Define “time to completion” ie:Identify time intervals: # of days from A → B → C → D → E → F → G → H, → I A→ I, or any combination Define “timely” ie: 30, 60, 90, 120, 180 days Define “kept appointment” & “not kept appointment” i.e. cancel vs reschedule vs no show |
Adherence to recommended diagnostic testing | Did not keep appointment rates Number of re-scheduled appointments |
Track for each scheduled appointment Analyzed at patient, provider, or practice level |
|
Stage at diagnosis | Stage 0–4 | TNM Classification of Malignant Tumors cancer staging criteria | *ACJJ Staging Manual55 |
Table 4.
Recommended common data elements for Process Metrics
Construct | Common Metrics | Core Data Elements | Analysis Strategies |
---|---|---|---|
Phase of cancer care that navigation program target | Phase of cancer care |
|
Check all that apply |
Caseload | Navigator Caseload | # of patients navigated per navigator Time spent per patient (minutes, hours) |
Define “caseload” within a specified timeframe (ie: weekly, monthly, annually) Consider accounting for time when defining caseload |
Communication | Mode of communication | Encounter type: in person, phone, letter Interpreter used (yes/no) |
Define “encounter” Track for each encounter Specify unit of analysis (per encounter, per patient) |
Barriers | Number of barriers Type of barriers |
Fitzpatrick scale PNRP Methods paper41 |
Define “barrier” Track for each encounter |
Actions | Number of actions Type of actions |
PNRP Methods paper41 | Define “action” Track for each encounter |
Table 2 displays recommendations for core data elements that may be used to create a set of common intermediate outcome metrics that fit within the constructs of screening outcomes, and potential analysis options for use in reporting. The first construct is Receipt of the screening test. Documenting the core data elements to measure this construct allows for the reporting of dichotomous outcomes metrics like “completion of screening test (yes/no)” or “timely completion of screening test (yes/no)” as well as the continuous outcome of “time to complete screening”. Either of these metrics may be reported using any or all of the common data points outlined in Table 2. This measure is limited when comparing programs with different eligibility criteria and follow up time periods. Thus, a more comparable construct to consider is Adherence to recommended screening, which requires of course that a screening guideline (such as the USPSTF) be stated explicitly. This common metric allows for programs to compare their adherence rates across different populations. Because the full benefit of screening on survival is dependent on the longitudinal use of “routine’ screening tests, there should be an emphasis towards documenting screening maintenance over time.
Table 3 displays recommendations for core data elements that may be used to create a set of common intermediate outcome metrics that fit within the constructs of diagnostic outcomes, and potential analysis options for use in reporting. Common measures for the construct of diagnostic resolution include the dichotomous outcome metrics of “completion of diagnostic resolution (yes/no)”, “timely completion of diagnostic resolution (yes/no)” and the continuous outcome of “time to complete diagnostic resolution.” These metrics may be reported using any or all of the common data elements outlined in Table 3. Resources and program intent will create variability in which data elements programs are interested in and capable of collecting. The priority should be to have an explicit and consistent definition of diagnostic resolution and to collect the date corresponding with that definition, as recommended by the PNRP.41 When the diagnostic resolution is a diagnosis of cancer, metrics such as stage at diagnosis are also important to record. Table 3 lists available options in care delivery that may constitute diagnostic resolution, which all programs should aspire if resources allow. Another recommended construct that is often omitted from program evaluation is Adherence to recommended testing, as determined by documentation of the type of diagnostic test performed in approaching diagnostic resolution. This common metric, completion of appropriate test, is another measure of quality to ensure populations have access to appropriate diagnostic testing.
Our third set of recommendations call for a minimal set of process data elements (Table 4). Process measures are intended to measure whether navigation was delivered as planned or designed. Without these details, replication of programs with successful outcomes is not possible. Knowledge of the specific components of a navigation program are necessary to apply lessons learned from one program to the next. The PED Working Group identified three distinct phases of PED where processes of navigation may differ:
Outreach/promotion (helping community understand the need and availability of cancer screening)
Support during clinical visits
Tracking and follow up after clinical visit completed
At a minimum, programs should document which phase(s) of PED their navigators address, as this captures broadly the types of activities involved in the navigation program.
In addition, we strongly recommend that programs document the number of patients navigated (over some specified time period: daily, weekly or monthly) and the time spent with individual patients. From this information measures of caseload may be created. Mode of communication and whether an interpreter was used in an encounter is another important process measure. Documenting the date of last navigator encounter ensures a way to attribute the screening outcome to navigation. For example, we wouldn’t want to attribute a screening outcome to navigation if there has been no contact with a navigator in the prior 12 months.
Considering that barriers to care are at the very center of the conceptual model of navigation, it is essential to measure them along with navigator activities, or actions, taken to address them. Creating an optimal set of patient-level barriers to care is challenging given the specific needs of diverse populations, as barriers in one community may be vastly different from barriers in another. Freund et al describe recommendations for barriers used by the PNRP41 and provide a framework for documenting navigation activities that would facilitate meaningful comparisons. The Native American Cancer Research Corporation (NACR) provides another example of documenting barriers and actions routinely used in their program.15 Finally, documenting healthcare utilization along the screening process is an alternative way of capturing benefits of navigation, such as reduction in rates of missed appointments.
Our fourth and final set of recommendations is related to data collection efforts. Data elements may be collected using patient self report, navigator logs, clinical data sources, and/or objective observation. At a minimum, programs should document their data source, given the limitations/strengths of these various sources. In a research context, it would be inappropriate to have navigators administer outcome assessments for their own patients as it would introduce potential bias. While it would be acceptable to have navigators document process measures, all programs should avoid using navigators to document clinical outcomes.
PN daily logs are an obvious source for process measures. Electronic programs can be used for those PN who have access to computers and/or the Internet. Tremendous effort should be made to ensure the layout consists of closed question format or checkboxes that address most prevalent responses with an “other” category that allows for text input. These lists or checkboxes should include space to document the amount of time the PN spent doing each task, or better yet checkboxes with time intervals.
Patient self-reported screening behaviors are often inaccurate.43 and how the questions are asked may influence the responses. However, if patient self report is used, phrasing of questions should be drawn from standardized, validated instruments such as the National Health Interview Survey (NHIS), the Behavioral Risk Factor Surveillance System (BRFSS), or the National Medical Care Expenditure Survey (NMEPS). Likewise, there are differences in the types of responses when such instruments are administered face-to-face, over the phone, completed by the patient, use of CADI (Computer Aided Design Instrument) systems and/or through the Internet.44
While objective observation methods of patients and navigators have been developed,45 most programs will not have the resources to utilize them. With federal mandates requiring transition to electronic medical records, there is tremendous opportunity to utilize objective clinical data sources to measure these outcomes and should be the standard for navigation programs to aspire. For example, electronic medical records may be queried for the presence of screening reports or as a means to complete certain data points, while electronic registration systems may be queried to report adherence outcomes for scheduled appointments.
Discussion
Patient navigation programs that target the prevention and early detection spectrum of care share similar goals, yet vary widely in how they document their success. Differences in program structure, population needs, outcomes of interest, and reported evaluation metrics make cross-study comparisons impossible. However, a review of the literature suggests that a common set of evaluation metrics relevant to multiple stakeholders can be developed. Based on a synthesis of existing navigation literature and expert consensus, we present here a set of four recommendations related to measuring and reporting PED navigation program success so that dissemination of the evidence may be used to delineate best practices in the design of care processes across diverse settings.
Our recommendations call for a core set of quality indicators that measure the intent of navigation – to bridge the critical disconnect between the discovery and delivery of life saving cancer care services. Knowledge of basic program characteristics is the starting point to contextualize comparisons between programs. While clinical outcome measures of quality (e.g., stage of diagnosis or mortality) are generally more difficult or not feasible to measure, we provide a framework of core data elements that may be used to report a common set of intermediate clinical metrics. Equally important, we provide recommendations for collecting and reporting process measures (activities performed while receiving care) which are the most frequently used quality indicators,46, 47 because they are sensitive, unambiguous, and easily measured.48–50 Our review of the literature highlights the lack of evidence linking these processes to clinical outcomes, making these data elements of high priority for future study as process measures should be associated to outcome measures for effective quality assessment. 51, 52
Priorities should focus on defining the needs and demographics of the target population, which in turn should drive the expected outcomes of the intervention. As long as programs explicitly state the context of their evaluation and choose from among the core set of data elements, meaningful comparisons among existing programs should be feasible. While methods for collecting these metrics will depend upon resources and existing infrastructure, programs should aspire for rigor with objective sources when possible. When objective electronic data are not available, sites need creativity to determine the best way to retrieve the information, either from manual chart abstraction or navigator documentation. These recommendations are a first step towards adopting a minimal dataset for PED navigation programs, as has been done by other population based approaches to improving quality care.53
Navigation is emerging as an expected "standard" for cancer programs,54 yet the literature has yet to provide consistent insight into activities or processes of navigation that are linked to favorable outcomes. We demonstrate here the growing body of knowledge regarding the impact of prevention and early detection navigation on cancer care would benefit from some thoughtful standardization. In keeping with recommendations from the 2001 IOM report to deliver patient-centered care that is timely, efficient and equitable3 it is imperative that we evaluate the ability of PED patient navigation programs to realize that potential. Only then can we “apply evidence to health care delivery” as recommended. The responsibility for the analysis and synthesis of this medical evidence falls on all of us involved in the delivery of these services.
References
- 1.Edwards BK, Ward E, Kohler BA, Eheman C, Zauber AG, Anderson RN, et al. Annual report to the nation on the status of cancer, 1975–2006, featuring colorectal cancer trends and impact of interventions (risk factors, screening, and treatment) to reduce future rates. Cancer. 116(3):544–73. doi: 10.1002/cncr.24760. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation…list_uids=19998273. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Berry DA, Cronin KA, Plevritis SK, Fryback DG, Clarke L, Zelen M, et al. Effect of screening and adjuvant therapy on mortality from breast cancer. N Engl J Med. 2005;353(17):1784–92. doi: 10.1056/NEJMoa050518. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=16251534. [DOI] [PubMed] [Google Scholar]
- 3.Crossing the Quality Chasm: A New Health System for the 21st Century. Institute of Medicine; 2001. [Google Scholar]
- 4.Elmore JG, Nakano CY, Linden HM, Reisch LM, Ayanian JZ, Larson EB. Racial inequities in the timing of breast cancer detection, diagnosis, and initiation of treatment. Med Care. 2005;43(2):141–8. doi: 10.1097/00005650-200502000-00007. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=15655427. [DOI] [PubMed] [Google Scholar]
- 5.Mandelblatt JS, Yabroff KR, Kerner JF. Equitable access to cancer services: A review of barriers to quality care. Cancer. 1999;86(11):2378–90. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=10590381. [PubMed] [Google Scholar]
- 6.A Summary of the American Cancer Society Report to the Nation: Cancer in the Poor. CA Cancer J Clin. 39:263–65. doi: 10.3322/canjclin.39.5.263. [DOI] [PubMed] [Google Scholar]
- 7.Wells KJ, Battaglia TA, Dudley DJ, Garcia R, Greene A, Calhoun E, et al. Patient navigation: state of the art or is it science? Cancer. 2008;113(8):1999–2010. doi: 10.1002/cncr.23815. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=18780320. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Freeman HP, Muth BJ, Kerner JF. Expanding access to cancer screening and clinical follow-up among the medically underserved. Cancer Pract. 1995;3(1):19–30. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=7704057. [PubMed] [Google Scholar]
- 9.Ell K, Padgett D, Vourlekis B, Nissly J, Pineda D, Sarabia O, et al. Abnormal mammogram follow-up: a pilot study women with low income. Cancer Pract. 2002;10(3):130–8. doi: 10.1046/j.1523-5394.2002.103009.x. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=11972567. [DOI] [PubMed] [Google Scholar]
- 10.Ell K, Vourlekis B, Lee PJ, Xie B. Patient navigation and case management following an abnormal mammogram: a randomized clinical trial. Prev Med. 2007;44(1):26–33. doi: 10.1016/j.ypmed.2006.08.001. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=16962652. [DOI] [PubMed] [Google Scholar]
- 11.Ell K, Vourlekis B, Muderspach L, Nissly J, Padgett D, Pineda D, et al. Abnormal cervical screen follow-up among low-income Latinas: Project SAFe. J Womens Health Gend Based Med. 2002;11(7):639–51. doi: 10.1089/152460902760360586. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=12396896. [DOI] [PubMed] [Google Scholar]
- 12.Battaglia TA, Roloff K, Posner MA, Freund KM. Improving follow-up to abnormal breast cancer screening in an urban population. A patient navigation intervention. Cancer. 2007;109(2 Suppl):359–67. doi: 10.1002/cncr.22354. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=17123275. [DOI] [PubMed] [Google Scholar]
- 13.Bastani R, Mojica CM, Berman BA, Ganz PA. Low-income women with abnormal breast findings: results of a randomized trial to increase rates of diagnostic resolution. Cancer Epidemiol Biomarkers Prev. 19(8):1927–36. doi: 10.1158/1055-9965.EPI-09-0481. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=20647406. [DOI] [PubMed] [Google Scholar]
- 14.Burhansstipanov L, Dignan MB, Schumacher A, Krebs LU, Alfonsi G, Apodaca CC. Breast screening navigator programs within three settings that assist underserved women. J Cancer Educ. 25(2):247–52. doi: 10.1007/s13187-010-0071-4. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=20300914. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Burhansstipanov L, Dignan MB, Wound DB, Tenney M, Vigil G. Native American recruitment into breast cancer screening: the NAWWA project. J Cancer Educ. 2000;15(1):28–32. doi: 10.1080/08858190009528649. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=10730800. [DOI] [PubMed] [Google Scholar]
- 16.American College of Surgeons Commission on Cancer. Cancer Program Standards (CPS) 2011 Project. Available at: http://www.facs.org/cancer/coc/cps2011.html. Accessed 12/6/2010., 2010.
- 17.Oluwole SF, Ali AO, Adu A, Blane BP, Barlow B, Oropeza R, et al. Impact of a cancer screening program on breast cancer stage at diagnosis in a medically underserved urban community. J Am Coll Surg. 2003;196(2):180–8. doi: 10.1016/S1072-7515(02)01765-9. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=12595043. [DOI] [PubMed] [Google Scholar]
- 18.Gabram SG, Lund MJ, Gardner J, Hatchett N, Bumpers HL, Okoli J, et al. Effects of an outreach and internal navigation program on breast cancer diagnosis in an urban cancer center with a large African-American population. Cancer. 2008;113(3):602–7. doi: 10.1002/cncr.23568. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=18613035. [DOI] [PubMed] [Google Scholar]
- 19.Dignan MB, Burhansstipanov L, Hariton J, Harjo L, Rattler T, Lee R, et al. A comparison of two Native American Navigator formats: face-to-face and telephone. Cancer Control. 2005;12(Suppl 2):28–33. doi: 10.1177/1073274805012004S05. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=16327748. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Phillips CE, Rothstein JD, Beaver K, Sherman BJ, Freund KM, Battaglia TA. Patient Navigation to Increase Mammography Screening Among Inner City Women. J Gen Intern Med. doi: 10.1007/s11606-010-1527-2. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=20931294. [DOI] [PMC free article] [PubMed]
- 21.Weber BE, Reilly BM. Enhancing mammography use in the inner city. A randomized trial of intensive case management. Arch Intern Med. 1997;157(20):2345–9. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=9361575. [PubMed] [Google Scholar]
- 22.Jandorf L, Gutierrez Y, Lopez J, Christie J, Itzkowitz SH. Use of a patient navigator to increase colorectal cancer screening in an urban neighborhood health clinic. J Urban Health. 2005;82(2):216–24. doi: 10.1093/jurban/jti046. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=15888638. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Lasser KE, Murillo J, Medlin E, Lisboa S, Valley-Shah L, Fletcher RH, et al. A multilevel intervention to promote colorectal cancer screening among community health center patients: results of a pilot study. BMC Fam Pract. 2009;10:37. doi: 10.1186/1471-2296-10-37. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=19480698. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Nash D, Azeez S, Vlahov D, Schori M. Evaluation of an intervention to increase screening colonoscopy in an urban public hospital setting. J Urban Health. 2006;83(2):231–43. doi: 10.1007/s11524-006-9029-6. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=16736372. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Chen LA, Santos S, Jandorf L, Christie J, Castillo A, Winkel G, et al. A program to enhance completion of screening colonoscopy among urban minorities. Clin Gastroenterol Hepatol. 2008;6(4):443–50. doi: 10.1016/j.cgh.2007.12.009. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=18304882. [DOI] [PubMed] [Google Scholar]
- 26.Paskett E, Tatum C, Rushing J, Michielutte R, Bell R, Long Foley K, et al. Randomized trial of an intervention to improve mammography utilization among a triracial rural population of women. J Natl Cancer Inst. 2006;98(17):1226–37. doi: 10.1093/jnci/djj333. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=16954475. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Han HR, Lee H, Kim MT, Kim KB. Tailored lay health worker intervention improves breast cancer screening outcomes in non-adherent Korean-American women. Health Educ Res. 2009;24(2):318–29. doi: 10.1093/her/cyn021. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=18463411. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Wang X, Fang C, Tan Y, Liu A, Ma GX. Evidence-based intervention to reduce access barriers to cervical cancer screening among underserved Chinese American women. J Womens Health (Larchmt) 19(3):463–9. doi: 10.1089/jwh.2009.1422. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=20156089. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Fernandez ME, Gonzales A, Tortolero-Luna G, Williams J, Saavedra-Embesi M, Chan W, et al. Effectiveness of Cultivando la Salud: a breast and cervical cancer screening promotion program for low-income Hispanic women. Am J Public Health. 2009;99(5):936–43. doi: 10.2105/AJPH.2008.136713. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=19299678. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Ford ME, Havstad S, Vernon SW, Davis SD, Kroll D, Lamerato L, et al. Enhancing adherence among older African American men enrolled in a longitudinal cancer screening trial. Gerontologist. 2006;46(4):545–50. doi: 10.1093/geront/46.4.545. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=16921009. [DOI] [PubMed] [Google Scholar]
- 31.Clark CR, Baril N, Kunicki M, Johnson N, Soukup J, Ferguson K, et al. Addressing social determinants of health to improve access to early breast cancer detection: results of the Boston REACH 2010 Breast and Cervical Cancer Coalition Women’s Health Demonstration Project. J Womens Health (Larchmt) 2009;18(5):677–90. doi: 10.1089/jwh.2008.0972. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=19445616. [DOI] [PubMed] [Google Scholar]
- 32.Percac-Lima S, Grant RW, Green AR, Ashburner JM, Gamba G, Oo S, et al. A culturally tailored navigator program for colorectal cancer screening in a community health center: a randomized, controlled trial. J Gen Intern Med. 2009;24(2):211–7. doi: 10.1007/s11606-008-0864-x. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=19067085. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Dietrich AJ, Tobin JN, Cassells A, Robinson CM, Greene MA, Sox CH, et al. Telephone care management to improve cancer screening among low-income women: a randomized, controlled trial. Ann Intern Med. 2006;144(8):563–71. doi: 10.7326/0003-4819-144-8-200604180-00006. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=16618953. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Christie J, Itzkowitz S, Lihau-Nkanza I, Castillo A, Redd W, Jandorf L. A randomized controlled trial using patient navigation to increase colonoscopy screening among low-income minorities. J Natl Med Assoc. 2008;100(3):278–84. doi: 10.1016/s0027-9684(15)31240-2. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=18390020. [DOI] [PubMed] [Google Scholar]
- 35.Ma GX, Shive S, Tan Y, Gao W, Rhee J, Park M, et al. Community-based colorectal cancer intervention in underserved Korean Americans. Cancer Epidemiol. 2009;33(5):381–6. doi: 10.1016/j.canep.2009.10.001. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=19914880. [DOI] [PubMed] [Google Scholar]
- 36.Myers RE, Hyslop T, Sifri R, Bittner-Fagan H, Katurakes NC, Cocroft J, et al. Tailored navigation in colorectal cancer screening. Med Care. 2008;46(9 Suppl 1):S123–31. doi: 10.1097/MLR.0b013e31817fdf46. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=18725824. [DOI] [PubMed] [Google Scholar]
- 37.Ferrante JM, Chen PH, Kim S. The effect of patient navigation on time to diagnosis, anxiety, and satisfaction in urban minority women with abnormal mammograms: a randomized controlled trial. J Urban Health. 2008;85(1):114–24. doi: 10.1007/s11524-007-9228-9. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=17906931. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Palmieri FM, DePeri ER, Mincey BA, Smith JA, Wen LK, Chewar DM, et al. Comprehensive diagnostic program for medically underserved women with abnormal breast screening evaluations in an urban population. Mayo Clin Proc. 2009;84(4):317–22. doi: 10.1016/S0025-6196(11)60539-9. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=19339648. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Psooy BJ, Schreuer D, Borgaonkar J, Caines JS. Patient navigation: improving timeliness in the diagnosis of breast abnormalities. Can Assoc Radiol J. 2004;55(3):145–50. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=15237774. [PubMed] [Google Scholar]
- 40.Fouad M, Wynn T, Martin M, Partridge E. Patient navigation pilot project: results from the Community Health Advisors in Action Program (CHAAP) Ethn Dis. 20(2):155–61. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=20503896. [PubMed] [Google Scholar]
- 41.Freund KM, Battaglia TA, Calhoun E, Dudley DJ, Fiscella K, Paskett E, et al. National Cancer Institute Patient Navigation Research Program: methods, protocol, and measures. Cancer. 2008;113(12):3391–9. doi: 10.1002/cncr.23960. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=18951521. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Lin CJ, Schwaderer KA, Morgenlander KH, Ricci EM, Hoffman L, Martz E, et al. Factors associated with patient navigators’ time spent on reducing barriers to cancer treatment. J Natl Med Assoc. 2008;100(11):1290–7. doi: 10.1016/s0027-9684(15)31507-8. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=19024225. [DOI] [PubMed] [Google Scholar]
- 43.Rauscher GH, Johnson TP, Cho YI, Walk JA. Accuracy of self-reported cancer-screening histories: a meta-analysis. Cancer Epidemiol Biomarkers Prev. 2008;17(4):748–57. doi: 10.1158/1055-9965.EPI-07-2629. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=18381468. [DOI] [PubMed] [Google Scholar]
- 44.Ponce NA, Lavarreda SA, Yen W, Brown ER, DiSogra C, Satter DE. The California Health Interview Survey 2001: translation of a major survey for California’s multiethnic population. Public Health Rep. 2004;119(4):388–95. doi: 10.1016/j.phr.2004.05.002. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=15219795. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Parker VA, Clark JA, Leyson J, Calhoun E, Carroll JK, Freund KM, et al. Patient navigation: development of a protocol for describing what navigators do. Health Serv Res. 45(2):514–31. doi: 10.1111/j.1475-6773.2009.01079.x. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=20132342. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Campbell SM, Roland MO, Shekelle PG, Cantrill JA, Buetow SA, Cragg DK. Development of review criteria for assessing the quality of management of stable angina, adult asthma, and non-insulin dependent diabetes mellitus in general practice. Qual Health Care. 1999;8(1):6–15. doi: 10.1136/qshc.8.1.6. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=10557672. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47.Donabedian A. Evaluating the quality of medical care. Milbank Mem Fund Q. 1966;44(suppl):166–206. [PubMed] [Google Scholar]
- 48.Hearnshaw HM, Harker RM, Cheater FM, Baker RH, Grimshaw GM. Expert consensus on the desirable characteristics of review criteria for improvement of health care quality. Qual Health Care. 2001;10(3):173–8. doi: 10.1136/qhc.0100173... Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=11533425. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49.Brook RH, McGlynn EA, Shekelle PG. Defining and measuring quality of care: a perspective from US researchers. Int J Qual Health Care. 2000;12(4):281–95. doi: 10.1093/intqhc/12.4.281. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=10985266. [DOI] [PubMed] [Google Scholar]
- 50.Crombie IK, Davies HT. Beyond health outcomes: the advantages of measuring process. J Eval Clin Pract. 1998;4(1):31–8. doi: 10.1046/j.1365-2753.1998.t01-1-00003.x. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=9524910. [DOI] [PubMed] [Google Scholar]
- 51.Rubin HR, Pronovost P, Diette GB. The advantages and disadvantages of process-based measures of health care quality. Int J Qual Health Care. 2001;13(6):469–74. doi: 10.1093/intqhc/13.6.469. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=11769749. [DOI] [PubMed] [Google Scholar]
- 52.Donabedian A. The quality of care: How can it be assessed? JAMA. 1988;260:1743–48. doi: 10.1001/jama.260.12.1743. [DOI] [PubMed] [Google Scholar]
- 53.Campbell HS, Ossip-Klein D, Bailey L, Saul J. Minimal dataset for quitlines: a best practice. Tob Control. 2007;16(Suppl 1):i16–20. doi: 10.1136/tc.2007.019976. Available from http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=18048624. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 54.American College of Surgeons Comission on Cancer. Cancer Program Accredidation. Availible at: http://www.facs.org/cancer/coc/sarguide.html accessed January 31, 2011.
- 55.American Joint Committee on Cancer. The A-JCC e-Staging Tool. Available at https://www.ajcce-staging.com/staging/view/cancerStage. Accessed January 3, 2011.