Abstract
Background
The English Cancer Patient Experience Survey (CPES) is a regularly conducted survey measuring the experience of cancer patients. We studied the survey's underlying structure using factor analysis to identify potential for improvements in reporting or questionnaire design.
Methods
Cancer Patient Experience Survey 2015 respondents (n = 71,186, response rate 66%) were split into two random subgroups. Using exploratory factor analysis (EFA) on the first subgroup, we identified the survey's latent structure. EFA was then applied to 12 sets of items. A first (“core”) set was formed by questions that applied to all participants. The subsequent sets contained the “core set” plus questions corresponding to specific care pathways/patient groups. We used confirmatory factor analysis (CFA) on the second data subgroup for cross‐validation.
Results
The EFA suggested that five latent factors underlie the survey's core questions. Analysis on the remaining 11 care pathway/patient group items also indicated the same five latent factors, although additional factors were present for questions applicable to patients with an overnight stay or those accessing specialist nursing. The five factors models had an excellent fit (comparative fit index = 0.95, root mean square error of approximation = 0.045 for core set of questions). Items loading on each factor generally corresponded to a specific section or subsection of the questionnaire. CFA findings were concordant with the EFA patterns.
Conclusion
The findings suggest five coherent underlying sub‐constructs relating to different aspects of cancer health care. The findings support the construction of evidence‐based composite indicators for different domains of experience and provide options for survey re‐design.
Keywords: cancer, factor analysis, health services administration, indicators, oncology, patient experience, psycho‐oncology, survey, surveys and questionnaires
The underlying structure of the Cancer Patient experience Survey comprises five major factors. These include shared decisions; care coordination; diagnostic care; timely tests; and aftercare. Our results support the current survey structure, but provide options to guide survey redesign.
1. BACKGROUND
Patient experience has been established as a distinct domain of quality of care, together with clinical effectiveness and patient safety. 1 , 2 , 3 Consequently, in recent decades modern healthcare systems conduct large patient surveys with nationwide coverage, whose findings are reported publicly for responsible/accountable organizations. Examples include the General Practice Patient Survey (GPPS) and the Adult Inpatient Survey in England, and the CAHPS surveys in the United States. 4 , 5 , 6 , 7 Although some such surveys encompass patients with any disease, some focus on the experience of patients with specific diseases. The English Cancer Patient Experience Survey (CPES) is an example of such a survey. 8 To 2020, there have been eight waves of this survey from 2010 onward, with another two waves being prepared.
Ideally, the psychometric properties of survey questionnaires are examined during the survey design process. Often, as was the case with the CPES, surveys are implemented prior to any psychometric evaluation. In such cases, factor analysis can provide insights with a number of potential uses. Factor analysis is a family of statistical techniques which identify underlying, latent, relationships among survey items, helping to identify the constructs underpinning a survey. Using factor analysis, survey questions which relate to the same underlying construct or domain of care can be grouped together. These domains of care could be used as the basis for performance management and public reporting conventions. Organizations may be classified on the basis of their performance within each domain, rather than, or in addition to, being classified on every item. Knowledge of these domains might help to more efficiently target quality improvement efforts, addressing the source of deficits in patient experience, rather than each particular aspect of experience measured by individual questions.
Having identified domains, results of factor analysis can also inform future questionnaire development. The number of questions relating to the domain, and the consistency of responses to questions within that domain, can help to inform whether further questions are needed or if there is potential for item removal within a domain. Although the approach has been applied to other patient experience surveys, 9 , 10 , 11 , 12 , 13 , 14 , 15 , 16 no prior study has used factor analysis to identify the underlying structure of CPES. We therefore aimed to elucidate the structure of the CPES survey using factor analysis.
2. METHODS
2.1. Data
We used data from 71,186 respondents to the National CPES 2015 (response rate 65.7%). Details of the survey and method of administration have been published elsewhere. 17 Briefly, the survey was mailed to all adult patients (aged 16 and over) discharged from a National Health Service hospital after inpatient or day case cancer‐related treatment during April–June 2015 following vital status checks at survey mail‐out (between 3 and 5 months after the sampling period).
The survey included 49 evaluative questions relating to aspects of patient experience (i.e., questions which ask patient to evaluate their care, which contrast with filter questions which often ask patients factual questions about their care to establish if a section of questions are relevant e.g., whether the patient has had an operation). It also includes questions about the patient (including age, gender, and ethnicity). Of the evaluative questions, seven have binary response options, 41 use a Likert scale with 3–7 response options, and one asks patients to rate their overall satisfaction between 0 and 10. Respondents were split randomly into two data sets, with one (N = 35,559) being used to establish the underlying structure of the data using exploratory factor analysis (EFA). The underlying factor structure was then confirmed using the second data set (N = 35,627) using confirmatory factor analysis (CFA).
2.2. Statistical analysis
2.2.1. Core questions
Of the 49 evaluative questions, 20 represented domains of care that were assumed to be relevant to all respondents (with the remaining questions being relevant only to certain groups of patients, such as those treated by chemotherapy or those in education or employment––see below). We excluded one of these questions (relating to access to clinical nurse specialists) from the core set as it acts both as a measure of experience and a filter question. Throughout this work, we refer to these 19 items as “core questions.” Despite the high overall response rate of the survey, only 27% (19,263) of respondents gave an informative response to all the core questions (i.e., answers such as ‘Don't know/can't remember where treated as missing). Restricting analyses to this subset of respondents would result in reduced precision and the potential for bias that can arise from pairwise deletion. 18 To counter this, we produced a single imputation of the missing responses using chained equations under the missing at random assumption. Predictive mean matching was used to maintain the interval nature of the data.
Reflecting the dichotomous and ordinal nature of the response options within CPES for most questions, we primarily employed categorical (polychoric) correlations within the EFA to avoid the attenuation of correlations between two categorical variables which can occur when Pearson correlations are used. We used linear (Pearson) correlations only for correlations involving the single 0–10 rating question. All correlations were computed using the psych package in R. 19 We first performed an unrestricted EFA to determine the number of factors to retain. We used two methods for determining this number of factors: the Kaiser criterion, which identifies and retains factors with eigenvalues greater than one 20 and the Cattell's scree test, which involves an examination of a plot of eigenvalues, the scree plot, for breaks or discontinuities. 21 Having identified the number of factors, we performed additional EFA restricted to the number of factors identified by either method. We applied oblique rotations (using the Promax rotation method as implemented by the psych package in R 19 ) when any of the two above methods indicated the retention of more than one factor with a view to explaining whether rotated models resulted in improved overall fit. These rotations lead to freely estimated inter‐factor correlations. 22 , 23 We use a cut‐off of 0.40 for the factor loadings. 24 Items with lower loadings were removed.
To account for the ordinal nature of responses, the factor structures from the EFA were examined within the CFA using structural equations models applying Satorra–Bentler adjustment to the standard errors and chi‐squared values. 22 , 23 We made use of the population error statistic root mean square error of approximation (RMSEA), the baseline comparison statistics comparative fit index (CFI) and Tucker Lewis index (TLI), and the standardised root mean squared residuals (SRMR) statistic . The following cut‐off values are presently recognized as indicative of good fit: RMSE < 0.07, CFI ≥ 0.95, TLI ≥ 0.95, SRMR < 0.08. 25 , 26 CFA was performed using the lavaan package in R. 27
The internal consistency reliability coefficients (Cronbach's alpha) for each factor derived from the EFA core model was computed using polychoric and Pearson correlations. The range of Cronbach's alpha coefficients in each factor when one question was left out was also calculated.
2.2.2. Questions relating to specific patient groups/care pathways
Unlike “core” questions that every patient could have answered, most questions applied only to specific patient groups (e.g., those in education or employment) or to those who have undergone specific care pathways (e.g., having been treated by chemotherapy or having had an overnight stay in hospital). When responses to these questions were missing, it usually reflected the lack of applicability of a specific care pathway (i.e., patients without a hospital stay , therefore they should not answer questions regarding their experience as inpatients), rather than reflecting a lack of response to an applicable question. For this reason, we did not impute responses to questions relating to specific patient groups/care pathways and aspects of care.
Following previous work examining key drivers of satisfaction, 28 we classified questions into 10 sets representing a specific patient group or care pathway, plus a further set including the question about access to clinical nurse specialists which was left out of the core set for analytic reasons (Appendix 1, Table A1). The above analysis for the core questions was repeated a further 11 times including responses to the core questions and responses to the questions applicable to the particular patient group/care pathway.
Analysis was performed using R 3.6.1. 29
3. RESULTS
3.1. Core questions
The scree plot for unrestricted EFA applied to the core set of 19 questions which were applicable to all respondents is shown in Figure 1 indicating that only one factor had an eigenvalue > 1, suggesting a single unidimensional underlying patient experience construct for the data. A restricted EFA model with a single factor resulted in factor loadings > 0.4 for 18 of the 19 questions considered (Table 1). Only the question on willingness to take part in cancer research (Q58) had a loading < 0.4. Applying this one factor model (after removing Q58) within a CFA found that, depending on the goodness‐of‐fit measures that were used, the model did not provide a good fit to the data (RMSEA = 0.081, CFI = 0.836, and TLI = 0.814 indicating an unacceptable fit, and SRMR=0.054 indicating an acceptable fit––against recommended normative threshold values of RMSE < 0.07, CFI ≥ 0.95, TLI ≥ 0.95, and SRMR < 0.08). 25 , 26
TABLE 1.
Question number | Core model |
---|---|
Factor 1 | |
Q2 | 0.457 |
Q6 | 0.518 |
Q7 | 0.708 |
Q8 | 0.530 |
Q9 | 0.577 |
Q10 | 0.609 |
Q11 | 0.613 |
Q12 | 0.761 |
Q13 | 0.736 |
Q14 | 0.779 |
Q15 | 0.705 |
Q16 | 0.787 |
Q49 | 0.644 |
Q52 | 0.534 |
Q54 | 0.697 |
Q55 | 0.439 |
Q56 | 0.670 |
Q58 | |
Q59 | 0.742 |
We therefore examined the model structure implied by examination of the scree plot to determine the number of factors which should be retained. The scree plot (Figure 1) did not display any clear break or discontinuity. We therefore chose to retain five factors, corresponding to the point where the outstep decline ends (after factor five) and reaches a very low level (at factors five and six). Applying a five factor restricted EFA to the 19 core questions identified factor loadings > 0.4 for all questions, except for Q58 (Table 2). In general, the factors correspond to a domain or subdomain of care as explicitly captured by a section or subsection of the survey questionnaire. The questions loading on each factor are as follows:
Factor 1: five items about treatment explanations and share decision‐making (Q12–Q16);
Factor 2: four items about care coordination and administration (Q52, Q54, Q56, and Q59);
Factor 3: five items about diagnosis (Q7–Q11);
Factor 4: two items about timeliness of investigations (Q2 and Q6);
Factor 5: two items about aftercare and support (Q49 and Q55).
TABLE 2.
Question number | Core model | ||||
---|---|---|---|---|---|
Factor 1 | Factor 2 | Factor 3 | Factor 4 | Factor 5 | |
Q2 | 0.79512 | ||||
Q6 | 0.79354 | ||||
Q7 | 0.43189 | ||||
Q8 | 0.58687 | ||||
Q9 | 0.80632 | ||||
Q10 | 0.69871 | ||||
Q11 | 0.51439 | ||||
Q12 | 0.55 | ||||
Q13 | 0.91428 | ||||
Q14 | 0.7041 | ||||
Q15 | 0.72134 | ||||
Q16 | 0.5222 | ||||
Q49 | 0.4723 | ||||
Q52 | 0.5632 | ||||
Q54 | 0.7609 | ||||
Q55 | 0.49 | ||||
Q56 | 0.7783 | ||||
Q58 | |||||
Q59 | 0.86681 |
Applying this five factor model within a CFA found that the model provided a good fit to the data for three of the four goodness‐of‐fit measures considered, with the fourth measure just below the acceptable threshold (RMSEA = 0.045 and SRMR = 0.029 indicating a good fit, CFI = 0.954 indicating an acceptable fit, and TLI = 0.944 just indicating an unacceptable fit).
The values of the Cronbach's alpha for each of the five factors fell within the acceptable 0.70 value with the exception of Factor 5, whose Cronbach's alpha was 0.60 (Appendix 1, Table A2). Deletion of one question from Factor 2 (“As far as you know, was your GP given enough information about your condition and the treatment you had at the hospital?”) led to an increase in Cronbach's alpha.
4. QUESTIONS RELATING TO SPECIFIC PATIENT GROUPS/CARE PATHWAYS
The scree plots for the unrestricted EFA models applied to the 11 sets of “patient group/pathway‐specific” questions (comprised of the core questions plus questions applicable to a particular patient group/care pathway) are shown in Appendix 2. When basing the number of factors to be retained on the basis of eigenvalues > 1 we retained only one factor in eight of the 11 patient group/pathway‐specific sets of questions. For the remaining three sets of questions, there are two eigenvalues > 1 (Appendix 2). As was the case with the core questions only model, in all 11 patient group/pathway‐specific EFA models restricted to one or two factors, the question on willingness to take part in cancer research (Q58) consistently had factor loadings < 0.4. This indicated that the question on willingness to take part in cancer research did not belong to either the core underlying construct of patient experience or the underlying construct of the additional patient group/pathway‐specific factor. For the 8/11 patient group/pathway‐specific sets of questions where (in restricted EFA) only one factor was retained, the noncore questions were all loaded (>0.4) onto this single factor, indicating they belonged to the core underlying construct of patient experience, with the exception of Q17 (“Were you given the name of a Clinical Nurse specialist?”). Where two factors were retained, the noncore questions all loaded onto a second factor defined by the noncore questions, namely;
Questions about support for people with cancer.
Questions about hospital stay.
Questions about support from health and social care services outside hospital.
One or more core questions were loaded onto the new factor. No cross loadings were observed on any of these models.
As with models restricted to the core questions, applying a one or two factor CFA as appropriate to the question sets, did not provide a good fit to the data (see Appendix 1, Table A3).
As with the core set of questions, the scree plots for the 11 sets of questions (comprised of the core questions plus questions applicable to a particular patient group/care pathway) did not display any clear break or discontinuity. Instead, we retained a number of factors such that all factors present in the core questions only model were retained. In nine of the 11 patient group/pathway‐specific sets of questions this was achieved by retaining five factors. In two cases, an additional factor was retained (resulting in six factor models, see Table 3) which related to
Questions about specialist nurse care.
Questions about hospital stay.
TABLE 3.
Core questions | Noncore questions | ||||
---|---|---|---|---|---|
Q no. | Synoptic form | Loading | Q no. | Synoptic form | Loading |
Factor 1—Shared decision‐making | |||||
12 | Treatment options explained | 0.550 a | 44 | Beforehand did you have all the information you needed for radiotx? | 0.534 h |
13 | Possible side effects of Tx explained/understandable? | 0.914 a | 45 | Were you given enough information about whether your radiotx was working? | 0.449 h |
14 | Practical advice and support in dealing with side effects of Tx | 0.704 a | 47 | Beforehand did you have all the information you needed for chemotx? | 0.703 i |
15 | Info on side effects of Tx in the longer term | 0.721 a | 48 | Were you given enough information about whether your chemotx was working? | 0.525 i |
16 | Were you involved in decisions about Tx as much as you wanted? | 0.522 a | |||
Factor 2—Care coordination and administration | |||||
52 | Did GP had enough info? | 0.563 a | 38 | Clear written information about what you should do or should not do after discharge | 0.699 f |
54 | Different people treating and caring for you work well together | 0.761 a | 39 | Were you told who to contact after leaving hospital if needed? | 0.683 f |
56 | Overall, how would you rate the administration of your care? | 0.778 a | 41 | Able to talk worries/fears | 0.412 g |
59 | Overall, how would you rate your care? | 0.867 a | 42 | Did they have the right documents? | 0.644 g |
53 | Did GP/practice nurse did their best? | 0.551 k | |||
Factor 3—Diagnostic process | |||||
7 | Were test results explained? | 0.432 a | |||
8 | Did you know you could bring family member? | 0.587 a | |||
9 | How do you feel about the way told Dx? | 0.806 a | |||
10 | Did you understand the explanation of your Dx? | 0.699 a | |||
11 | When told Dx, were you given written info about the type of cancer? | 0.514 a | |||
Factor 4—Timeliness of investigations | |||||
2 | Times you saw your GP before going to hospital about cancer | 0.795 a | |||
6 | How did you feel about the length of time you had to wait for test? | 0.794 a | |||
Factor 5—Aftercare and support | |||||
49 | Did family/someone close get all info to help care for you at home? | 0.472 a | 17 | Were you given the name of a specialist nurse? | 0.461 b |
55 | Have you been given a care plan? | 0.489 a | 20 | Hospital gave enough info about support groups | 0.782 d |
21 | Hospital staff discuss about impact of cancer on work/education | 0.638 d | |||
22 | Hospital staff gave info about financial help | 0.869 d | |||
23 | Hospital staff gave info on free prescriptions | 0.757 d | |||
50 | During your tx were you given enough care and support from health or social services? | 0.912 j | |||
51 | After your tx were you given enough care and support from health or social services? | 1.045 j | |||
Factor 6—Specialist nursing | |||||
18 | Was it easy to contact your specialist nurse? | 0.819 c | |||
19 | Did the specialist nurse give answers you could understand? | 0.802 c | |||
Factor 7—Hospital stay | |||||
29 | Confidence and trust in doctors | 0.481 e | |||
31 | Confidence and trust in nurses | 0.914 e | |||
32 | In your opinion, were there enough nurses? | 0.693 e | |||
34 | Enough privacy to discuss | 0.586 e | |||
36 | Did they do everything to control pain? | 0.664 e | |||
37 | Treated with respect and dignity | 0.946 e |
From model with the 19 core questions applicable to all patients.
From five factor model with core questions and the question about access to clinical nurse specialists.
From six factor model with core questions and questions applicable to patients with access to specialist nurse.
From five factor model with core questions and questions applicable to patients with recent hospital care.
From six factor model with core questions and questions applicable to patients with recent hospital stay.
From six factor model with core questions and questions applicable to patients with recent hospital stay.
From five factor model with core questions and questions applicable to patients with recent outpatient or day case appointments.
From five factor model with core questions and questions applicable to patients treated by radiotherapy.
From five factor model with core questions and questions applicable to patients treated by chemotherapy.
From five factor model with core questions and questions applicable to patients who received support from health and social care services.
From five factor model with core questions and question applicable to patients with recent outpatient appointments.
In CFA, these five and six factor models (Appendix 1, Table A3) were found to provide a good/acceptable fit to the data according to RMSEA and SRMR (RMSEA range 0.042–0.058 and SRMR range 0.028–0.056). The CFI and TLI statistics for these models were closed to achieving, or achieved, an acceptable fit (CFI range 0.931–0.954 and TLI range 0.919–0.944).
5. DISCUSSION
5.1. Summary of findings
We have applied exploratory and confirmatory factor analyses to the responses to the English CPES. We found that the core set of questions which applied to all patients, and many questions which applied only to a subset of patients were dominated by a single underlying factor (as indicated by factor eigenvalues > 1). However, this single factor did not provide a good description of the data (according to goodness‐of‐fit metrics), thus implying a more complex underlying structure. Visual inspection of scree plots implied that five underlying factors can describe the experiences of patients captured by the core questions applicable to all patients. These included: shared decision‐making; care coordination and administration; the diagnostic process; timeliness of investigations; and aftercare and support. Many questions applicable to specific subsets of patients also fitted within the latter five underlying domains, but additional factors were required for specialist nursing and for hospital stay. These domains of care provide a good description of the data (according to goodness‐of‐fit metrics). Furthermore, they largely fitted with the existing structure of the survey and, in light of the data presented here, represent a reasonable target for public reporting of data and performance improvement.
5.2. Comparisons with the literature
Previous work has examined the underlying structure of other nationwide patient experience surveys including HCAHPS and GPPS, 30 , 31 but this is the first time that this approach has been applied to an established nationwide experience of cancer patients. It has been long recognized that patient experience varies greatly by patient sociodemographic characteristics, including age, sex, socioeconomic status, and ethnicity. 32 , 33 Furthermore, for CPESs, cancer site/type is strongly associated with ratings of experience, above and beyond adjustment for other patient‐level variables. 32 , 34 Future work should address the question whether the underlying structure of CPES may vary by patient group.
5.3. Strengths and limitations
We used a large sample, which allowed for precise estimation of underlying factors. We have only used data from a single year, however the survey has been conducted seven times between 2010 and 2019. During this period, there have been only small changes made to the wording of survey items. However, the number and type of question have remained largely the same, and the overall structure has remained consistent with the same sections covering the various stages of the care pathway. Therefore, the findings are likely to be generalizable across survey waves.
We note that in general core questions relating to the different factors tend to be placed in close proximity to each other within the questionnaire. While this can be useful to the patient, it is possible that this proximity influenced the factor structure we observed. However, when we also considered the questions applicable to certain patient groups or care pathways we found similar questions loading on to the same factor which were placed at some distance within the questionnaire. This would not be expected if proximity was the driving force behind the observed factor and thus provides further support for the 5/6 factor structure we propose.
5.4. Implications
Our results support the current structure of the survey which in general covers the range of aspects of care and patient experience which are relevant to cancer patients. The survey seems to capture the experience of patient groups defined by different care pathways and services equally well. Furthermore, the results indicate that although factual questions (such as about participating in research, Q58) could be successfully included in care experience questionnaires, it is important to recognize that these do not, on the basis of results presented here, represent aspects of patient experience per se.
This analysis can be used as the basis of supporting the construct of a number of composite indicators to summarize hospital performance with respect to cancer patient experience. For example, such composites, might target organizational performance across aspects of care experience relating to the five underlying domains/factors identified (i.e., shared decision‐making; care coordination and administration; the diagnostic process; timeliness of investigations; and aftercare and support). Alongside consideration of the drivers of satisfaction with care, such composites may help users of the survey to more easily relate to study findings and prioritize bundles of actions and interventions targeting specific composite domains (we explore this further below). This could, in principle, help to increase the reliability of organizational‐level scores, which are known to represent a limitation of question‐based scores of the CPES survey, though this needs to be explored directly in further empirical research.
It is worth noting that while factor analysis can provide evidence about patterns of responses, it tells us little about the relative importance of the various aspects of care. If an overall summary score were to be derived, various weighting schemes could be applied. All questions or domains of care could be weighted equally, though this implies they have equal importance. Alternatively, policy‐based weights may be employed reflecting an external view of the importance of different domains of care. A third option is to employ a key drivers analysis which empirically examines the importance of survey items to survey responders using their associations with a global evaluation item. Such a key drivers analysis has been carried out for a number of surveys, including CPES. 28 , 35 , 36 , 37 Many of these use a selection of individual questions, but others use domain scores, which can be based on factor analyses.
The CPES is a survey with a relatively large number of questions. As such there may be some desire to shorten the questionnaire to reduce burden on responding patients. The high internal consistency of Factor 1 (shared decision‐making) and Factor 3 (diagnostic process) indicates the potential for item removal. In contrast, the low internal consistency of Factor 5 (aftercare and support) indicates that there may be benefit in additional questions in this area. While factor analysis can help to identify potential questions for removal (for example by identifying domains of experience survey by a large number of questions) it should be noted that factor analysis is not considered sufficient for such purposes. 22 , 23 First, removing a question from a survey could be detrimental to its content validity. Furthermore, weak loadings might be the result of sampling error, although this is unlikely to be an issue in our study context, given the large sample size. As a consequence, replication of factor analytic models is critical for scale development.
6. CONCLUSION
The underlying structure of the CPES corresponds to five major aspects of care experience and pathways of cancer patients. The findings support the current survey design, though they also provide potential options to guide survey redesign, and have potential to inform the way the survey findings might optimally be reported, and improvement efforts targeted.
CONFLICT OF INTEREST
The authors report grants from MacMillan Cancer Support, during the conduct of the study. GA and GL have acted as academic consultants providing methodological advice to NHS England Insight team regarding the Cancer Patient Experience Survey during 2015‐17.
ETHICS APPROVAL
This research used anonymized data that are available for bona fide research by the UK Data Archive. As such no ethics approval of such research is required.
PATIENT CONSENT FOR PUBLICATION
This work is a secondary analysis of anonymous data collected as part of a national survey programme. Each participant was provided with a detailed patient information leaflet and return of a questionnaire implied consent.
ACKNOWLEDGMENTS
The authors thank the NHS England as commissioner of the 2015 Cancer Patient Experience Survey, Quality Health as data collector and publisher of the data set, all NHS Acute Trusts in England for provision of data samples, and all patients who responded to the survey.
APPENDIX 1.
TABLE A1.
Synoptic form | Patient group number | Patient group name |
---|---|---|
Q17. Were you given the name of a CNS | 1 | Question excluded from core set due to analytic reasons |
Q18. Was it easy to contact your CNS | 2 | Patients with access to specialist nurse |
Q19. Did the CNS give answers you could understand? | ||
Q20. Hospital gave enough info about support groups | 3 | Patients with recent hospital care |
Q21. Hospital staff discuss about impact of cancer on work/education | ||
Q22. Hospital staff gave info about financial help | ||
Q23. Hospital staff gave info on free prescriptions | ||
Q26. How operation had gone explained afterward | 4 | Patients with recent operation |
Q28. Did doctors speak of you as if you were not there? | 5 | Patients with recent hospital stay |
Q30. Were family members able to speak to doctor? | ||
Q33. Were you asked how you want to be called? | ||
Q35. Able to discuss worries/fears | ||
Q29. Confidence and trust in doctors | ||
Q31. Confidence and trust in nurses | ||
Q32. In your opinion, were there enough nurses? | ||
Q34. Enough privacy to discuss | ||
Q36. Did they do everything to control pain? | ||
Q37. Treated with respect and dignity | ||
Q38. Clear written information about what you should (not) do after discharge | ||
Q39. Were you told who to contact after leaving hospital if needed | ||
Q41. Able to talk worries/fears | 6 | Patients with recent outpatient or day case appointments |
Q42. Did they have the right documents? | ||
Q44. Beforehand did you have all the information you needed for radiotx | 7 | Patients treated by radiotherapy |
Q45. Were you given enough information about whether your radiotx was working? | ||
Q47. Beforehand did you have all the information you needed for chemotx | 8 | Patients treated by chemotherapy |
Q48. Were you given enough information about whether your chemotx was working? | ||
Q50. During your tx were you given enough care/support from health/social services? | 9 | Patients who received support from health and social care services |
Q51. After your tx were you given enough care/support from health/social services? | ||
Q53. Did GP/practice nurse did their best? | 10 | Patients accessing primary care post‐discharge |
Q57. Overall, how do you feel about the length of time to wait for clinics and appointments? | 11 | Patients with recent outpatient appointments |
TABLE A2.
Cronbach's alpha coefficient | Range of alphas if individual items deleted | |
---|---|---|
Factor 1: Shared decision‐making (five items) |
0.90 | 0.87–0.89 |
Factor 2: Care coordination and administration (four items) | 0.74 | 0.61–0.79 |
Factor 3: Diagnostic process (five items) |
0.81 | 0.76–0.79 |
Factor 4: Timeliness of investigations (two items) |
0.79 | — |
Factor 5: Aftercare and support (two items) |
0.60 | — |
TABLE A3.
Model | One/two factors | Five/six factors | ||||||
---|---|---|---|---|---|---|---|---|
RMSEA | CFI | TLI | SRMR | RMSEA | CFI | TLI | SRMR | |
Core | 0.081 | 0.836 | 0.814 | 0.054 | 0.045 | 0.9540 | 0.944 | 0.0290 |
Core + CNS name | 0.081 | 0.836 | 0.814 | 0.054 | 0.044 | 0.9510 | 0.941 | 0.0290 |
Core + CNS | 0.079 | 0.8150 | 0.793 | 0.0550 | 0.04 | 0.953 | 0.943 | 0.028 |
Core + support for cancer patients | 0.073 | 0.872 | 0.857 | 0.049 | 0.046 | 0.946 | 0.938 | 0.033 |
Core + operation | 0.079 | 0.838 | 0.818 | 0.053 | 0.045 | 0.954 | 0.944 | 0.029 |
Core + overnight stay | 0.058 | 0.882 | 0.871 | 0.046 | 0.042 | 0.935 | 0.926 | 0.034 |
Core + outpatient/day case | 0.075 | 0.837 | 0.818 | 0.052 | 0.045 | 0.946 | 0.936 | 0.031 |
Core + radiotx | 0.079 | 0.824 | 0.804 | 0.056 | 0.047 | 0.941 | 0.93 | 0.034 |
Core + chemotx | 0.075 | 0.844 | 0.826 | 0.052 | 0.045 | 0.947 | 0.937 | 0.031 |
Core + health/social services | 0.095 | 0.837 | 0.814 | 0.073 | 0.058 | 0.931 | 0.919 | 0.056 |
Core + care from GP | 0.082 | 0.826 | 0.804 | 0.055 | 0.049 | 0.943 | 0.931 | 0.031 |
Core + overall care | 0.078 | 0.832 | 0.811 | 0.053 | 0.045 | 0.954 | 0.944 | 0.029 |
Abbreviations: CFI, comparative fit index; RMSEA, root mean square error of approximation.
APPENDIX 2.
Gomez‐Cano M, Lyratzopoulos G, Campbell JL, Elliot M, Abel G. The underlying structure of the English Cancer Patient Experience Survey: Factor analysis to support survey reporting and design. Cancer Med.2022;11:3–20. doi: 10.1002/cam4.4325
Funding information
This research was funded by Macmillan Cancer Support (grant number 5995414). GL is funded by a Cancer Research UK Advanced Clinician Scientist Fellowship award (grant number C18081/A18180). The work presented here represents that of the authors and does not necessarily represent those of the funding organizations.
Contributor Information
Mayam Gomez‐Cano, Email: m.c.gomez-cano@exeter.ac.uk.
Gary A. Abel, Email: g.a.abel@exeter.ac.uk.
DATA AVAILABILITY STATEMENT
Data used in this study are available via the UK Data Archive, Study Number (SN) 8163.
REFERENCES
- 1. Locock L, Graham C, King J, et al. Health services and delivery research. In Understanding How Front‐Line Staff Use Patient Experience Data for Service Improvement: An Exploratory Case Study Evaluation. NIHR Journals Library; 2020. [PubMed] [Google Scholar]
- 2. Llanwarne NR, Abel GA, Elliott MN, et al. Relationship between clinical quality and patient experience: analysis of data from the English Quality and Outcomes Framework and the National GP Patient Survey. Ann Fam Med. 2013;11(5):467‐472. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3. Campbell J. Patients' experience of primary care: James Mackenzie Lecture 2017. Br J Gen Pract. 2019;69(678):38‐39. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4. Care Quality Commission . Adult inpatient survey 2018. Accessed November 11, 2021. https://nhssurveys.org/surveys/survey/02‐adults‐inpatients/year/2018/
- 5. Centers for Medicare and Medicaid Services Consumer Assessment of Healthcare Providers and Systems (CAHPS). Accessed September 07, 2018. https://www.cms.gov/Research‐Statistics‐Data‐and‐Systems/Research/CAHPS/
- 6. Agency for Healthcare Research and Quality . CAHPS Cancer Care Survey. 2017. Accessed December 02, 2021. https://www.ahrq.gov/cahps/surveys‐guidance/cancer/index.html
- 7. NHS England . GP Patient Survey. 2016. Accessed December 02, 2021. https://www.england.nhs.uk/statistics/2016/07/07/gp‐patient‐survey‐2015‐16/
- 8. Quality Health . National Cancer Experience Survey 2016. Accessed November 11, 2021. https://www.gov.uk/government/statistics/national‐cancer‐patient‐experience‐survey‐2016‐national‐data
- 9. Campbell J, Smith P, Nissen S, Bower P, Elliott M, Roland M. The GP Patient Survey for use in primary care in the National Health Service in the UK—development and psychometric characteristics. BMC Family Practice. 2009;10(1):57. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10. Marshall GN, Morales LS, Elliott M, Spritzer K, Hays RD. Confirmatory factor analysis of the Consumer Assessment of Health Plans Study (CAHPS) 1.0 core survey. Psychol Assess. 2001;13(2):216‐229. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11. Hays RD, Martino S, Brown JA, et al. Evaluation of a care coordination measure for the Consumer Assessment of Healthcare Providers and Systems (CAHPS®) medicare survey. Med Care Res Rev. 2014;71(2):192‐202. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12. Lee Hargraves J, Hays RD, Cleary PD. Psychometric properties of the consumer assessment of health plans study (CAHPS®) 2.0 adult core survey. Health Serv Res. 2003;38(6p1):1509‐1528. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13. Solomon LS, Hays RD, Zaslavsky AM, Ding L, Cleary PD. Psychometric properties of a group‐level Consumer Assessment of Health Plans Study (CAHPS) instrument. Med Care. 2005;43(1):53‐60. [PubMed] [Google Scholar]
- 14. O'Malley AJ, Zaslavsky AM, Hays RD, Hepner KA, Keller S, Cleary PD. Exploratory factor analyses of the CAHPS® hospital pilot survey responses across and within medical, surgical, and obstetric services. Health Serv Res. 2005;40(6p2):2078‐2095. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15. Dyer N, Sorra JS, Smith SA, Cleary P, Hays R. Psychometric properties of the Consumer Assessment of Healthcare Providers and Systems (CAHPS®) clinician and group adult visit survey. Med Care. 2012;50(Suppl):S28‐S34. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16. Ramsay J, Campbell JL, Schroter S, Green J, Roland M. The General Practice Assessment Survey (GPAS): tests of data quality and measurement properties. Fam Pract. 2000;17(5):372‐379. [DOI] [PubMed] [Google Scholar]
- 17. Quality Health . National Cancer Patient Experience Survey 2015. National Results Summary. Accessed December 02, 2021. https://www.ncpes.co.uk/wp‐content/uploads/2020/06/National‐Cancer‐Patient‐Experience‐Survey‐2015‐National‐Report.pdf
- 18. Shi D, Lee T, Fairchild AJ, Maydeu‐Olivares A. Fitting ordinal factor analysis models with missing data: a comparison between pairwise deletion and multiple imputation. Educ Psychol Measur. 2020;80(1):41‐66. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19. Revelle W. psych: procedures for psychological, psychometric, and personality research. In R package version 1812. 2018. Accessed November 11, 2021. https://cran.r‐project.org/web/packages/psych/index.html [Google Scholar]
- 20. Kaiser HF. The application of electronic computers to factor analysis. Educ Psychol Measur. 1960;20(1):141‐151. [Google Scholar]
- 21. Hayton JC, Allen DG, Scarpello V. Factor retention decisions in exploratory factor analysis: a tutorial on parallel analysis. Organ Res Methods. 2004;7(2):191‐205. [Google Scholar]
- 22. Flora DB, Flake JK. The purpose and practice of exploratory and confirmatory factor analysis in psychological research: decisions for scale development and validation. Can J Behav Sci. 2017;49(2):78. [Google Scholar]
- 23. Bandalos DL, Finney SJ. Exploratory and confirmatory. In Hancock GR, Stapleton LM, Mueller RO, eds. The Reviewer's Guide to Quantitative Methods in the Social Sciences New York. Taylor and Francis; 2018;93‐114. [Google Scholar]
- 24. Howard MC. A review of exploratory factor analysis decisions and overview of current practices: what we are doing and how can we improve? Int J Hum‐Comput Int. 2016;32(1):51‐62. [Google Scholar]
- 25. Hooper D, Coughlan J, Mullen MR. Structural equation modelling: guidelines for determining model fit. Electron J Bus Res Methods. 2008;6(1):53‐60. [Google Scholar]
- 26. Lt HU, Bentler PM. Cutoff criteria for fit indexes in covariance structure analysis: conventional criteria versus new alternatives. Struct Equ Modeling. 1999;6(1):1–55. [Google Scholar]
- 27. Rosseel Y. lavaan: an R package for structural equation modeling. J Stat Softw. 2012;48:1‐36. [Google Scholar]
- 28. Gomez‐Cano M, Lyratzopoulos G, Abel GA. Patient experience drivers of overall satisfaction with care in cancer patients: evidence from responders to the English Cancer Patient Experience Survey. J Patient Exp. 2020;7(5):758–765. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29. R Core Team . R: A Language and Environment for Statistical Computing. R Foundation for Statistical Computing; 2019. [Google Scholar]
- 30. Cefalu MS, Elliott MN, Setodji CM, Cleary PD, Hays RD. Hospital quality indicators are not unidimensional: a reanalysis of Lieberthal and Comer. Health Serv Res. 2019;54(2):502‐508. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31. Campbell J, Smith P, Nissen S, Bower P, Elliott M, Roland M. The GP Patient Survey for use in primary care in the National Health Service in the UK—development and psychometric characteristics. BMC Family Practice. 2009;10(1):57. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32. Abel G, Saunders C, Lyratzopoulos G. General Practice Profiles for Cancer: how variable are the practices and can we reliably judge their diagnostic performance? Eur J Cancer Care. 2015;24:72. [Google Scholar]
- 33. Lyratzopoulos G, Elliott M, Barbiere JM, et al. Understanding ethnic and other socio‐demographic differences in patient experience of primary care: evidence from the English General Practice Patient Survey. BMJ Qual Saf. 2012;21(1):21‐29. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34. El Turabi A, Abel GA, Roland M, Lyratzopoulos G. Variation in reported experience of involvement in cancer treatment decision making: evidence from the National Cancer Patient Experience Survey. Br J Cancer. 2013;109:780‐787. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35. Quigley DD, Elliott MN, Farley DO, Burkhart Q, Skootsky SA, Hays RD. Specialties differ in which aspects of doctor communication predict overall physician ratings. J Gen Intern Med. 2014;29(3):447‐454. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36. Collins RL, Haas A, Haviland AM, Elliott MN. What matters most to whom. Med Care. 2017;55(11):940‐947. [DOI] [PubMed] [Google Scholar]
- 37. Elliott MN, Kanouse DE, Edwards CA, Hilborne LH. Components of care vary in importance for overall patient‐reported experience by type of hospitalization. Med Care. 2009;47(8):842‐849. [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
Data used in this study are available via the UK Data Archive, Study Number (SN) 8163.