Abstract
Background & objectives
The comparative uses of different types of patient experience (PE) feedback as data within quality improvement (QI) are poorly understood. This paper reviews what types are currently available and categorizes them by their characteristics in order to better understand their roles in QI.
Methods
A scoping review of types of feedback currently available to hospital staff in the UK was undertaken. This comprised academic database searches for “measures of PE outcomes” (2000‐2016), and grey literature and websites for all types of “PE feedback” potentially available (2005‐2016). Through an iterative consensus process, we developed a list of characteristics and used this to present categories of similar types.
Main results
The scoping review returned 37 feedback types. A list of 12 characteristics was developed and applied, enabling identification of 4 categories that help understand potential use within QI—(1) Hospital‐initiated (validated) quantitative surveys: for example the NHS Adult Inpatient Survey; (2) Patient‐initiated qualitative feedback: for example complaints or twitter comments; (3) Hospital‐initiated qualitative feedback: for example Experience Based Co‐Design; (4) Other: for example Friends & Family Test. Of those routinely collected, few elicit “ready‐to‐use” data and those that do elicit data most suitable for measuring accountability, not for informing ward‐based improvement. Guidance does exist for linking collection of feedback to QI for some feedback types in Category 3 but these types are not routinely used.
Conclusion
If feedback is to be used more frequently within QI, more attention must be paid to obtaining and making available the most appropriate types.
Keywords: accountability, consensus exercise, feedback, Friends & Family Test, hospitals, NHS Inpatient Survey, patient experience, quality improvement, wards
1. INTRODUCTION
The importance of listening and responding to the voices of patients and carers as a means of supporting high quality and safe care in hospital settings has been strongly advocated.1, 2, 3 The use of patient experience (PE) feedback as a data tool within quality improvement (QI) is a seemingly logical step as evidenced by a systematic review4 into how different types of feedback have been used in QI, and a more discursive piece on PE feedback as measurement data.5 Both reveal an immature field of study providing more questions than answers, relating to what feedback to collect, and how and when, and then how to use feedback to inform and measure QI. This lack of certainty has not prevented its collection: we know that much feedback is collected but is not used.6 We also know that when staff are presented with feedback and encouraged to use it for QI, they are faced with a complexity of social and logistical barriers.7 In a period of shrinking resources when capability for data management and PE is notably stretched,8 we must improve our knowledge of the potential roles of different types of feedback in QI so that resources can be directed appropriately.
We note a number of specific uncertainties surrounding the potential for different types of PE feedback to be used in QI. Most apparent is the debate over the comparative value of quantitative and qualitative types. In the systematic review of uses of PE types,4 quantitative surveys were revealed to be the most frequently collected type of PE data (often mandated) but the least acceptable to health‐care teams with respect to use within QI, considered by many to reflect PE as conceived externally, rather than providing useful information for improvement. Conversely, teams like more qualitative types of feedback (ie, more in‐depth accounts of individuals’ experiences) that are perceived to more closely portray their patients’ concerns. However, reviews have also identified an apparent sense of nervousness amongst hospital teams surrounding the use of qualitative data as it is regarded as time‐consuming to collect4 and difficult to interpret without bias.5
Second is the critique directed at the data source currently most readily available in England—the mandated Friends and Family Test (FFT). Whilst proponents argue it offers timely, continuous and local‐level data ripe for use in QI at many levels,9 others10 suggest that problems of validity and representation make comparisons between time and space impossible, and the lack of qualitative detail with which to contextualize results, mean this tool is not fit for purpose. Third is the growing interest in utilizing types of data that are not collected specifically by an organization for improvement, but are available for use should organizations wish. So, there is interest in utilizing complaints as data,11, 12 as well as online reviews13 but our understanding of these sources is embryonic, and some argue that organizations will struggle to engage with such sources if they did not seek, sanction or solicit them.14 Finally, there is the arrival of frameworks that link feedback collection to a QI change process in which involvement of staff and patients is generally high and localized. These include methods such as Experience‐Based Co‐Design for which evidence of impact is growing, but less evidence of cost‐effectiveness of this resource‐intensive technique relative to other QI processes is available.15
One response to this growing appreciation of the different types of PE feedback potentially available has been to suggest that health‐care staff should mix different (and multiple) types of PE data to triangulate and obtain the most comprehensive information for improvement.5 However, this ambition is also debated with one critique focusing on the loss of meaning that can arise if rich, “untameable” qualitative data are “aggregated” and “triangulated” with quantifiable data in an attempt to arrive at more tractable issues.16 Instead, it is argued that “softer” less quantitative data serve a different purpose—disrupting assumptions rather than counting occurrences. The assertion that qualitative PE feedback provides additional (not just supporting) insights necessary for understanding aspects of PE not possible to elicit through quantitative surveys has been made by others.17, 18 These are the “relational” aspects so important to concepts of PE (eg, how were you treated?) as opposed to more transactional components (eg, was a service provided on time?) targeted by surveys. With this in mind, triangulation and aggregation concepts provide no neat solutions to handling different types.
Amidst these uncertainties then, there have already been some attempts to distinguish between feedback types according to potential purpose. In 2013, an evidence scan19 outlined a wide range of PE feedback types available, from quantitative surveys to qualitative patient stories, and characterized them by their ability to generalize (quantitative types) or describe (qualitative types). Subsequently, there have been two reviews of quantitative PE surveys available worldwide—one20 assesses them for utility arguing that their primary use is for “high‐stake purposes” such as benchmarking, hospital rankings and securing funding. The other21 concludes similarly and also summarizes why they are not suitable for informing local (eg, ward, clinical team, disease group) improvement initiatives: they do not provide locally attributable data and they lack nuance and detail. However, the role of surveys in local‐level improvement has not been discounted altogether: it has been proposed that some surveys, if designed and supported to allow local interpretation and timely processing, could be used to monitor changes within local improvement process over time, in addition to benchmarking functions.22, 23
Indeed, the distinction between different uses of data within improvement is not a new one and in 1997, “The 3 Faces of Performance Measurement” were outlined: data used for accountability (outcome measurements of interest to external parties, eg, funders and regulators), data for improvement process (detailed information to aid identification of problems, opportunities for change and monitoring of success) and data for research (generating universal knowledge).24 We wish to build on all of these existing distinctions in order to guide improved utilization of the plethora of different PE feedback types that can now exist so that health‐care staff can make more informed choices about what can be achieved by engaging with them as data in improvement. To do so, we document the following three‐stage process that used UK hospitals as a case study, with which to understand types of data arising from different types of PE feedback:
A scoping review of all types of PE feedback currently available to hospital staff in the UK that builds on previous reviews of surveys to include other feedback available.
Development of a list of characteristics that we believe to be important in understanding potential use within QI that consolidates what is already known combined with our own research experience of improving quality of care.
Use of these characteristics to define types of feedback identified in our scoping review into distinct categories that can begin to inform policymakers, researchers and those responsible for collecting and using PE feedback, of their potential comparative uses.
Whilst we use NHS hospitals in the UK as a case study, we anticipate our characteristics list and categories to be relevant to types of PE feedback that arise in different hospitals elsewhere, but also as categories that could be used as a starting point to consider feedback in any health and social care setting where staff are considering how feedback can be used to improve services.
2. METHODS
2.1. A scoping review of sources of PE feedback in the UK
Between Spring and Autumn 2016, we conducted a scoping review comprising academic databases, grey literature databases and websites, and supported this with our own knowledge from the field and that of our study steering group. We also hand‐searched citations contained within returned documents. We identified surveys from the existing reviews18, 19 and then conducted our own search of academic databases to update and focus on the UK only. We used grey literature and websites to identify other types of PE feedback that we knew, because of their non‐validated status, were not likely to be found in academic journals, but more likely to be discussed in “guidance” documents and commentaries. We adopted a scoping review method, and not a systematic review, because flexibility of search terms within grey literature was paramount to enable as wide a range of PE feedback to be returned. Comprehensiveness of sources available in the UK, whilst important, was secondary to our aim of developing a characterization system and categories that we anticipate could be applicable to other types as they emerge. We were informed by a five‐step framework for conducting scoping reviews 25 as shown in Table 1.
Table 1.
Identifying the research question | “What sources of PE feedback are currently available to hospital staff in the UK?” |
Identifying relevant studies |
Search of academic databases (Medline; Cinahl Plus; Amed; Scopus; Web of Science; Psych INFO; ProQuest Hospital collection) using terms: ‘patient experience’*’patient’’ outcome assessment (healthcare)”, measures*. Timeframe: 2000‐2016. Search of grey literature (Google, Google Scholar, Grey Literature Database, Royal College of Nursing database, Care Quality Commission (CQC), Collaborations for Leadership in Applied Health Research and Care (CLAHRC), Health Foundation, HealthTalk.org, iWantGreatCare, Health Watch, Kings Fund, NHS England, NHS Institute for Innovation & Improvement, NHS Surveys, Mumsnet, Patients Like Me, Patient Experience Portal, Patient Experience Network, Care Opinion, Picker Institute, Scottish Government, World Health Organization) using terms ‘patient experience feedback within the NHS’, ‘patient experience feedback of hospital care’, ‘NHS use of patient experience feedback of hospital healthcare’, ‘improving patient experience’, ‘patient experience toolkit’. These were subsequently adapted to suit different ways organizations use terms. Timeframe: 2005‐2016—narrower than for academic databases due to high volume of returns. NB Different terms were used for academic databases than those for grey literature because of the different content likely to be returned through each route. |
Study selection |
Inclusion criteria: any sources of feedback relating to patient experience of hospital care; patient or carer perspective; for use in UK acute hospital setting. Exclusion criteria: sources of feedback relating to patient experience of specific aspects of quality such as safety, clinical outcomes, person centred care, performance of individual clinicians or health‐care staff, treatment/condition specific experiences; not patient or carer perspective; not secondary care; those under 18 y; for use outside UK. |
Charting the data |
The search returned 37 different types of PE feedback for which we immediately created 3 broad categories that were informed by our general understanding of the way feedback varied. This enabled the results to be displayed in 4 separate tables to aid comparison: Appendices S1a: (17 surveys), 1b (12 patient‐initiated feedback) and 1c (7 hospital‐initiated qualitative feedback). We found that 2 types of feedback did not fit well in any and placed these in a 4th Table as Appendix S1d (other). This was deemed a reasonably objective task and was therefore performed by one researcher (RP) with two additional researchers confirming these categories. |
Collating, summarizing and reporting the results | Our categorization exercise, which is detailed below, fulfils this stage. |
2.2. Developing a list of “defining characteristics”
We established a consensus team to develop a list of 12 key descriptive characteristics to help understand the role of different feedback types in QI. This list is provided in Table 2. The team comprised the PI (Professor Psychology of Healthcare), 4 health service researchers (1 psychologist, 2 social scientists and 1 sociologist), 2 design researchers (concerned with presentation and usability of patient feedback), 1 health‐care improvement specialist and 1 patient involvement facilitator. The list developed iteratively through the following stages:
The PI first used evidence referred to above, combined with own knowledge of QI and PE to produce an initial list of nine characteristics and presented this to the consensus team.
The consensus team then added a further four characteristics to make 13.
One researcher (RP) attempted to use this list to characterize six of the types returned through the review finding that twelve of the characteristics worked effectively and only one did not so this was removed. This was “whether the feedback only related to specific patient groups” which was not possible to ascertain from descriptions of the types.
This list of 12 was then presented to the study steering group (comprising 4 lead researchers, 6 staff and 6 patient representatives from 3 hospital trusts, 2 national PE advisors) to ensure it made sense beyond the consensus team. This process led to clarification of the definitions and potential variability (character options) of each characteristic as listed in Table 2.
Table 2.
Characteristics of PE feedback | Character options |
---|---|
Nature of data obtained from feedback | |
Type |
|
Level of applicability |
|
Evidence for validity (applies to surveys only) |
|
Timing of feedback |
|
Mode of feedback collection | |
Mode of feedback collection |
|
Availability of feedback | |
Requirement for feedback |
|
Supporting hospital systems |
|
Timeliness of feedback availability to service |
|
Regularity of feedback |
|
Perspective captured | |
Who initiatives feedback? |
|
Who provides feedback? |
|
Defined role in QI | |
Extent of the defined role |
|
2.3. Assigning “characteristics”
These characteristics were then applied by RP to all returned types of PE, which enabled a definition of each type to be summarized into four broad categories within our Appendix S1 tables and described below. These were checked by two other members of the team before finalizing.
3. FINDINGS
We used our characteristics list to further understand and subdivide the feedback types within our initial four broad categories. This process also enabled us to provide more indicative titles for the categories than those we used as Appendix S1 titles. The categories and subcategories are shown in Table 3a‐d. The distinctions that we make between them are now described, highlighting potential implications for role within improving PE.
Table 3.
(A) | |
---|---|
(a) Mandated | (b)Voluntary |
Hospital‐level:
The NHS Adult Inpatient Survey (England)26, 27, 28 Scottish Inpatient Patient Experience Survey29 Inpatient Patient Experience Survey (NI)30 Any level: Your NHS Patient Experience Survey (Wales)31 Service or speciality: NHS A&E Survey (England)28, 32 NHS Maternity Services Survey (England)28 Scottish Maternity Care Survey33 |
Hospital‐level:
Hospital Care & Discharge34 Any level: PPE 1535 OxPIE36 Newcastle Satisfaction with Nursing Scale37 VOICE38 Service or speciality: PEECH 39 ICE Questionnaire40 New Models Study41 Urgent Care System42 Patient carer diary43 |
(B) | |
---|---|
(a) Formal hospital system | (b) No hospital system |
Complaints44, 45, 46, 47
Liaison Service concerns46, 47, 48, 49, 50 Hospital‐supported feedback cards Hospital‐supported websites: NHS Choices51 iWantGreatCare52 Facebook set up by ward/hospital Care Opinion (if adopted)54 |
Compliments and Thank you cards17
Websites: Mumsnet53 Google reviews of hospitals Facebook (generally) Care Opinion (if not adopted)54 Other websites |
3.1. Four categories of types of PE feedback
Seventeen types of feedback fitted into the first category “Hospital‐initiated quantitative surveys”.26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43 Common to almost all of these is that data are predominantly quantitative, initiated by hospitals, targeting patients and not carers, with a significant delay (due to processing) in providing information back to the organization. However, closer inspection reveals a distinction between those that are mandated for high‐level organization use (either whole organization or for whole A&E or whole maternity departments) at regular but infrequent intervals, and those that are offered as voluntary tools for use as and when an organization decides. The former most clearly exhibit accountability features—providing organizational‐level data, within parameters defined and initiated by the organization, validated to make generalizations, comparisons (between organizations, or over time) when conducted for large samples. They are long—having over 70 items—and so require significant processing. On the other hand, with the exception of one (Hospital Care & Discharge34), the voluntary surveys can be applied at any level, at a timing to suit, or are especially designed for use within a local service or speciality (eg, ICE 40), without prescribing regularity. Unlike the mandatory surveys, only some are clearly validated. Many of these are significantly shorter—around 20 items. Potentially, these more flexible surveys that elicit local‐level information offer more scope for informing, or monitoring local improvement of PE. Only one survey (Your NHS Patient Survey Wales31) does not conform neatly to this subdivision. This survey is strongly recommended for use, not mandated, and is designed for use at any level. This implies more flexibility, and that perhaps it has been designed to inform or monitor local improvements, as well as to provide accountability.
We call the second category “Patient‐initiated qualitative feedback” and include 12 feedback types here44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54 that exhibit common traits: they provide qualitative data, applicable to any level of the organization, are initiated by patients on an ad hoc basis (whenever they choose to) and the feedback is available to the organization quickly (referred to as in real‐time). The concept of validity is not applicable because all data are provided on a case‐by‐case basis. Within this category, the significant distinction is between those types that are formally supported and those that are not. For those that are, this could be because they are mandated to do so (complaints44, 45, 46, 47; concerns46, 47, 48, 49, 50; NHS choices51), or because they choose to adopt a system (set up a ward‐based Facebook page or buy into iWantGreatCare52) to organize their feedback. Other types have no supporting system in place and include informal feedback (compliments, Thank you cards) that is received but not perceived of as data requiring attention or processing. We include a caveat here because some hospitals could have more formal systems for handling these (we know anecdotally that this happens) but this is not widely acknowledged or articulated as a process. This subcategory also includes websites external to the organization (eg, Facebook, Twitter, Mumsnet,53 Google reviews) where patients/carers may upload feedback, but there is no guarantee this will be viewed by hospital staff. Other less well‐known sites could also exist on the Internet. Care opinion54 currently spans both subcategories: it is offered as a formal system of data management for a fee if hospitals choose to adopt this. If not formally adopted, the platform can still be used by patients to upload feedback that may or may not be viewed by the hospital.
In summary, this category offers a different kind of “data” than that offered in Category 1, and therefore has a potentially different role within QI. In Category 1, feedback offers evidence‐based scope for use in benchmarking and monitoring of organizational trends. Category 2 feedback provides more local‐level information that would not be valid for use in that way. It exhibits some characteristics (nuance, specificity) that suggest potential use within local QI processes especially problem identification. Currently, however, feedback within this category is presented largely on a case‐by‐case basis and not as collated data ready to use. This makes its proposed role as a data source more tentative than the surveys of Category 1, and we return to this issue in the Discussion.
We name the third category “Hospital‐initiated qualitative feedback,” and this includes six types of feedback55, 56, 57, 58, 59, 60 with some common, defining features: feedback is predominantly qualitative and can be collected for any level of service by a variety of methods with varying degree of prescription in this regard. Interviews are common but focus groups, observation and shadowing all feature here. All types elicit rich data that takes time to process. All have a defined role within QI, albeit to a varied extent. Feedback collection is initiated by staff but in striking contrast to surveys (which cover issues deemed important to organizations about their service delivery) qualitative methods are used in ways designed to explore patient/carers' experiences in an open unrestricted manner—content being determined by what is important to them. Unlike data elicited from feedback types in Category 2, methods used within Category 3 are designed to elicit rich and collated data sets ready to use.
PE types within this category also come with varying levels of guidance for linking collection of feedback to QI techniques. Some advocate linking feedback directly into continuous learning process (action research within Patient Journeys55; metrics collected elsewhere are used to track progress within Kinda Magic56; mainstream QI approaches such as PDSA are used in 15 Steps and Always Events58, 59). EBCD/AEBCD57 recommends collecting qualitative feedback of impact to assess perceptions of how the service has changed and also suggests collecting other measures about the change, for example cost‐savings to a service. Three also stand out for the way they use feedback as data for problem definition. Within EBCD/AEBCD and Always Events, feedback is interpreted together with staff and patients/carers in a process of co‐design so that contextual meaning informed by those who work in the service can be added. Patient Stories60 offer less in the way of prescribed QI process than the others, focusing only on one aspect—the presentation of the story to people who can potentially make changes as a result (often Boards). Subsequent change techniques are implied but not defined.
Finally, we identify a fourth category of miscellaneous “Other” with two types of feedback61, 62: FFT61 and HowRWe62 do not fit any of the above three categories. They are both surveys that hospitals can initiate, asking standardized questions, but unlike those surveys in Category 1, they are not designed to capture large quantities of data (lots of questions) infrequently, but instead they are very short and designed to be used more frequently, potentially providing a more continuous flow of PE feedback. In both cases, their responses are requested on a scale (positive to negative), allowing qualitative comments to be added, and they can both be applied to any type of health‐care setting. This is where their similarities end however. The nature of the questioning is very different. FFT only has one question: How likely are you to recommend our service to friends and family if they needed similar care or treatment? This is noticeably different to HowRWe which uses its very short design (four questions) to ask specific things about kind treatment, listening, promptness and organization, more akin to the content of surveys in Category 1. The data arising from the HowRWe standardized questions are validated to provide comparable data over time and between areas, whereas the data arising from the FFT standardized question are not. Potentially then, the HowRWe tool has more obvious potential for measurement and monitoring of trends within QI overtime and between areas than FFT but it is FFT that is mandatory in England whereas HowRWe is a voluntary tool and therefore much less widespread. The qualitative comments arising from both of these tools can be likened to the data arising from the feedback types included in Category 2: qualitative and context‐specific therefore holding potential to be used to guide local‐level improvement, however, they are not provided as collated data ready to be used, and the steps to enable them to be used as data are not specified.
4. DISCUSSION
In this paper, we have responded to demands made for PE feedback to be used more effectively in QI4, 5, 6 by conducting a scoping review and characterization exercise of different PE feedback types, to highlight their various potential roles. This builds on recent attempts to distinguish between roles depending on the nature of data produced,19, 20, 21, 22, 23 which we believe can be helpfully linked to grounding concepts of measurement within QI.24 Our scoping review identified 37 different types of PE feedback “on offer” to staff within UK hospitals. Using a consensus exercise, we drafted a list of characteristics that we believed to be important indicators of potential roles for each type. Using these characteristics to assess each type, we arrived at four distinct categories that we named: “Hospital‐initiated quantitative surveys”; “Patient‐initiated qualitative feedback”; “Hospital‐initiated qualitative feedback”; and “Other.” We have described above the nature of each of these categories with reference to roles within QI. In addition, we make the following observations:
4.1. Mandated sources currently provide limited value within QI
Of the mandated PE feedback types available, none of these would appear immediately suitable for informing and monitoring local improvement process (eg, ward level). Mandated feedback currently comprises quantitative survey data (the national inpatient surveys for whole organizations,26, 27, 28, 29, 30, 31 A&E28, 32 and maternity departments28, 33), complaints and liaison service data44, 45, 46, 47, 48, 49, 50 one form of online feedback (NHS Choices51) and the Friends and Family Test results.61 The large quantitative surveys—many validated for representativeness with large samples—serve an accountability purpose but do not provide locally relevant data that are accessible to those who need it, in a timely manner21, 63 that would be required for informing and monitoring QI process. The FFT quantitative test9 appears to seek to address both requirements. Its qualitative, locally applicable information could, in principle, be used to inform what needs to improve but this proposal is fiercely questioned,10 described as a laudable ambition thwarted by an overemphasis on achieving acceptable response rates at the expense of considering and utilizing the qualitative comments effectively. The lack of standardization with respect to administration of its single quantitative question has also cast doubts on its suitability as a monitoring tool over time. As the surveys are not providing everything required for QI, there is increasing interest in the use of other mandated feedback—from complaints and liaison services—by coding and theming into data sets.11 However, there are challenges to these proposals,12 relating to system practicalities (collation of case‐by‐case complaints), the nature of the story told (complex and difficult to code) and availability (often infrequent and inconsistent in style). In short, seen from a QI perspective, mandatory PE data (national surveys, FFT, complaints/concerns and NHS Choices), currently appear to offer little ready‐to‐use data, with respect to informing and monitoring local PE improvement.
4.2. Other types of feedback offer potential
Other, non‐mandated types of feedback are available should hospitals wish to use them but an understanding of their potential uses is in relative infancy. Hospitals could use voluntary surveys of Category 1—these offer more granular data and can be used more flexibly if analytical capability exists.23 They could use the patient‐initiated qualitative types (eg, complaints, comments, social media reviews) which, due to their spontaneous nature, arguably tap into patients own concerns more readily than anything requested from the health‐care organizations themselves. Indeed, systems for harnessing such sources are emerging and include the development of dedicated websites for encouraging and organizing this feedback (eg, Care Opinion,54 iWantGreatCare,52 dedicated Facebook pages for some hospitals/wards). There is also an emerging interest in harnessing “the cloud of patient experience” from social media that could exist independently of any hospital systems just because the public use these platforms to discuss their hospital services (eg, Twitter, Facebook, Google).64 Finally, Category 3 (hospital‐initiated qualitative types) offer something else again—processes for not only collection of feedback but for the development of shared meanings about issues of importance, in a cooperative approach involving patients and staff, and in many cases, processes of action and reflection as a response to identified priorities.
4.3. Beyond the concept of PE metrics
In traditional QI theory,24 all three faces of measurement (accountability, improvement and universal knowledge) imply the need for data that contain objective metrics that can be tracked for changes over time. The nature of Category 3 provides some important insights about how PE feedback can be conceptualized within a QI process. Feedback collected through Category 3 methods elicits rich, open‐ended information, and staff are supported to engage meaningfully with these concerns as part of a continual learning process about how services can be improved as a result. In these approaches, feedback is not so much viewed as static metrics (objective data) but, linked to the concept of “soft intelligence”,16 it is viewed as a mechanism for disrupting staff assumptions and making space for patient/carer perspectives. For example, within two techniques (EBCD/AEBCD and Always Events), the aim is for staff and patients/carers to develop shared meanings from the feedback and within EBCD/AEBCD specifically; this is described as a co‐design and co‐creation process involving techniques that aid critical, collective reflection.65 Clearly, within this broader interpretation of feedback, the concept of triangulation introduced above5 as the basis through which to view multiple sources together is limited: some feedback provides metrics that may be identifiable in different sources, but some sources (eg, Category 3) will not elicit these metrics and are performing a function more akin to soft intelligence.16
4.4. Limitations
Due to the flexible approach taken to search terms, our scoping review may not have revealed all potential feedback types available in UK hospitals. Also, due to the sometimes subjective nature of these search terms, a repeat exercise by others may not yield exactly the same results. Our characterization exercise was based on best understanding at the time—at a different point in time, other characteristics could have been chosen. For example, we included a characteristic called “Supported by hospital system” to refer to whether or not the hospital invited, encouraged or organized this feedback. Since then, the term SSS (sanctioned, solicited, sought) has been introduced14 to distinguish between online feedback that organizations support, and that which exists independently, and such a term may have provided more clarity if we had been able to use it. When categorizing based on our characteristics, some subjective decisions were also made. In some cases, there was ambiguity and we used our characteristics list as a sensitizing framework rather than an absolute.
5. CONCLUSION
Our scoping review has confirmed that there are many different types of PE feedback available, or potentially available, within UK hospitals that appear to reflect a very worthwhile ambition to ensure patient/carer voice influences improvement of services. However, our characterization and categorization study has revealed that within these types, there are currently no “ready‐to‐use” data sets for informing and monitoring improvements to PE, apart from mandated data relating to high‐level organizational trends. Many decisions therefore have to be made about the extent to which hospitals engage with different types of PE feedback for their improvement initiatives. The categories we have introduced highlight the important differences to consider which could aid these decisions in hospitals and potentially in other health‐care environments. We know hospital teams are already struggling to handle feedback that they are mandated to collect8; therefore, informed decisions with respect to these options are crucial. To support this, we propose further analysis and conceptual development of the role of PE feedback within QI, and that the categories we present in this paper can be used to inform this process.
CONFLICT OF INTEREST
All authors declare that they have no conflict of interest.
AUTHORS’ CONTRIBUTIONS
CM oversaw study design and coordinated, conceptualized and wrote the paper with the support of RP, and will act as overall guarantor. RP and LH conducted the scoping review, coordinating input from Kathryn Melling. RL and LS convened and led the consensus team that developed the characteristics list and categories used to sort the findings of the scoping review. All authors contributed to the conceptualization and writing of the manuscript drafts, revised the manuscript critically for important intellectual content and gave final approval of the version to be published.
ETHICAL APPROVAL
No ethical approvals were required to conduct this exercise.
Supporting information
ACKNOWLEDGEMENTS
The authors thank Kathryn Melling, previously of Bradford Institute for Health Research, for conducting a proportion of searches for the scoping review (academic databases for patient experience measures).
Marsh C, Peacock R, Sheard L, Hughes L, Lawton R. Patient experience feedback in UK hospitals: What types are available and what are their potential roles in quality improvement (QI)? Health Expect. 2019;22:317–326. 10.1111/hex.12885
Funding information
This research was funded by the NIHR Health Services and Delivery programme (Ref: 14/156/32). The research was supported by the NIHR CLAHRC Yorkshire and Humber (www.clahrc‐yh.nihr.ac.uk). The views expressed are those of the authors and not necessarily those of the NHS, the NIHR or the Department of Health.
REFERENCES
- 1. Francis R. Report of the Mid Staffordshire NHS Foundation Trust Public Inquiry. London, UK: The Stationary Office; 2013. [Google Scholar]
- 2. Keogh B. Review into the quality of the care and treatment provided by 14 hospitals in England. London, UK: The Stationary Office; 2013. [Google Scholar]
- 3. Berwick D. A promise to learn – a commitment to act: improving the safety of patients in England. London, UK: Department of Health; 2013. [Google Scholar]
- 4. Gleeson H, Calderon A, Swami V, Deighton J, Wolpert M, Edbrooke‐Childs J. Systematic review of approaches to using patient experience data for quality improvement in healthcare settings. BMJ Open. 2016;6(8):e011907. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5. LaVela S, Gallan A. Evaluation and measurement of patient experience. PXJ. 2014;1(1):28‐36. [Google Scholar]
- 6. Coulter A, Locock L, Ziebland S, Calabrese J. Collecting data on patient experience is not enough: they must be used to improve care. BMJ. 2014;348:g2225. [DOI] [PubMed] [Google Scholar]
- 7. Sheard L, Marsh C, O'Hara J, Armitage G, Wright J, Lawton R. The patient feedback response framework – understanding why UK hospital staff find it difficult to make improvements based on patient feedback: a qualitative study. Soc Sci Med. 2017;178:19‐27. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8. Gilbert D, Making GN. Making sense and making use of patient experience data. Membership Engagement Services&In Health Associates2015.
- 9. Churchill N. Making the most of feedback. 2018. https://www.england.nhs.uk/blog/making-the-most-of-feedback/. Accessed February 1, 2018.
- 10. Robert G. Friends and family test should no longer be mandatory. BMJ. 2018;360:k367. [DOI] [PubMed] [Google Scholar]
- 11. Gillespie A, Reader TW. The healthcare complaints analysis tool: development and reliability testing of a method for service monitoring and organisational learning. BMJ Qual Saf. 2016;25:937‐946. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12. de Vos MS, Hamming JF, Marang‐van de Mheen PJ. The problem with using patient complaints for improvement. BMJ Qual Saf. 2018;27:758‐762. [DOI] [PubMed] [Google Scholar]
- 13. Griffiths A, Leaver MP. Wisdom of patients: predicting the quality of care using aggregated patient feedback. BMJ Qual Saf. 2018;27:110‐118. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14. Dudhwala F, Boylan A‐M, Williams V, Powell J. VIEWPOINT: What counts as online patient feedback, and for whom? Digital Health. 2017;3:1‐3. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15. Donetto S, Tsianakas V, Robert G. Using Experience‐based Co‐design to Improve the Quality of Healthcare: Mapping Where We are Now and Establishing Future Directions. London, UK: King’s College London; 2014. [Google Scholar]
- 16. Martin GP, McKeeb L, Dixon‐Woods M. Beyond metrics? Utilizing ‘soft intelligence’ for healthcare quality and safety. Soc Sci Med. 2015;142:19‐26. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17. Entwistle V, Firnigl D, Ryan M, Francis J, Kinghorn P. Which experiences of health care delivery matter to service users and why? A critical interpretive synthesis and conceptual map. J Health Serv Res Policy. 2012;17(2):70‐78. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18. Van der Vleuten CP, Schuwirth LW. Assessing professional competence: from methods to programmes. Med Educ. 2005;39(3):309‐317. [DOI] [PubMed] [Google Scholar]
- 19. de Silva D. Measuring Patient Experience. London, UK: The Health Foundation; No.18; 2013. [Google Scholar]
- 20. Beattie M, Murphy DJ, Atherton I, Lauder W. Instruments to measure patient experience of healthcare quality in hospitals: a systematic review. Syst Rev. 2015;4:97. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21. Edwards KJ, Walker K, Duff J. Instruments to measure the inpatient hospital experience: A literature review. PXJ. 2015;2(2):77‐85. [Google Scholar]
- 22. Gibbons EJ, Graham C, Flott KM, Jenkinson C, Fitzpatrick R. Developing approaches to the collection and use of evidence of patient experience below the level of national surveys. PXJ. 2016;3(1):92‐100. [Google Scholar]
- 23. Flott KM, Graham C, Darzi A, Mayer E. Can we use patient‐reported feedback to drive change? The challenges of using patient‐reported feedback and how they might be addressed. BMJ Qual Saf Published Online First. 2017;26:502‐507. [DOI] [PubMed] [Google Scholar]
- 24. Solberg LI, Mosser G, Mcdonald S. The three faces of performance measurement: improvement, accountability & research. Jt Comm J Qual Improv. 1997;22(3):135‐147. [DOI] [PubMed] [Google Scholar]
- 25. Arksey H, O'Malley L. Scoping studies: towards a methodological framework. Int J Soc Res Methodol. 2007;1:19‐32. [Google Scholar]
- 26. Coulter A, Cleary PD. Patients' experiences with hospital care in five countries. Health Aff (Millwood). 2001;20(3):244‐252. [DOI] [PubMed] [Google Scholar]
- 27. Reeves R, West E, Barron D. Facilitated patient experience feedback can improve nursing care: a pilot study for a phase III cluster randomised controlled trial. BMC Health Serv Res. 2013;13:259. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28. NHS England . NHS Surveys. 2016http://www.nhssurveys.org/surveys. Accessed April 13, 2016.
- 29. Scottish Government . 2016 Inpatient Experience Survey. 2016. http://www.gov.scot/Topics/Statistics/Browse/Health/InpatientSurvey. Accessed April 13, 2016.
- 30. Department of Health, Social Services and Public Safety (Northern Ireland) . Inpatient Patient Experience Survey. 2016. https://www.health-ni.gov.uk/sites/default/files/publications/dhssps/inpatient-patient-experience-survey-2014.pdf. Accessed April 13, 2016.
- 31. NHS Wales . Your NHS Patient Experience Survey. 2016. http://www.powysthb.wales.nhs.uk/sitesplus/documents/1145/Q%26S_Item_6.3_Appendix%203_Healthcare%. Accessed April 14, 2016.
- 32. Bos N, Sizmur S, Graham C, van Stel HF. The accident and emergency department questionnaire: a measure for patients' experiences in the accident and emergency department. BMJ Qual Saf. 2013;22(2):139‐146. [DOI] [PubMed] [Google Scholar]
- 33. Scottish Government . Maternity Care Survey. 2015http://www.gov.scot/Topics/Statistics/Browse/Health/maternitysurvey. Accessed April 14, 2016.
- 34. Jones D, Lester C. Hospital care and discharge: patients' and carers' opinions. Age Ageing. 1994;23(2):91‐96. [DOI] [PubMed] [Google Scholar]
- 35. Jenkinson C, Coulter A, Bruster S. The Picker Patient Experience Questionnaire: development and validation using data from in‐patient surveys in five countries. Int J Qual Health Care. 2002;14(5):353‐358. [DOI] [PubMed] [Google Scholar]
- 36. Hewitson P, Skew A, Graham C, Jenkinson C, Coulter A. People with limiting long‐term conditions report poorer experiences and more problems with hospital care. BMC Health Services Research. 2014;14:33. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37. Thomas LH, McColl E, Priest J, Bond S, Boys RJ. Newcastle satisfaction with nursing scales: an instrument for quality assessments of nursing care. Qual Health Care. 1996;5(2):67‐72. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38. Evans J, Rose D, Flach C, et al. VOICE: developing a new measure of service users' perceptions of inpatient care, using a participatory methodology. J Ment Health. 2012;21(1):57‐71. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39. Murrells T, Robert G, Adams M, Morrow E, Maben J. "Measuring relational aspects of hospital care in England with the 'Patient Evaluation of Emotional Care during Hospitalisation' (PEECH) survey questionnaire. BMJ Open. 2013;3(1):e002211. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40. Rattray J, Johnstone M, Wildsmith JA. The intensive care experience: development of the ICE questionnaire. J Adv Nurs. 2004;47(1):64‐73. [DOI] [PubMed] [Google Scholar]
- 41. Fitzpatrick R, Graham C, Gibbons E, King J, Flott K, Jenkinson C. Development of new models for collection and use of patient experience information in the NHS – PRP 070/0074, Final Report. Oxford, UK: Picker Institute Europe; 2014.
- 42. Knowles E, O’Cathain A, Nicholl J. Patients’ experiences and views of an emergency and urgent care system. Health Expect. 2012;15(1):78‐86. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43. Baker R, Preston C, Cheater F, Hearnshaw H. Measuring patients’ attitudes to care across the primary/secondary interface: the development of the patient carer diary. Qual Health Care. 1999;8:154‐160. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44. NHS England . How to complain to the NHS. 2015. http://www.nhs.uk/NHSEngland/complaints-and-feedback/Pages/nhs-complaints.aspx. Accessed May 3, 2016.
- 45. Scottish Government . Complain about an NHS Service. 2016. https://www.mygov.scot/nhs-complaints. Accessed May 3, 2016.
- 46. NHS Wales . NHS Complaints. 2016. http://www.wales.nhs.uk/ourservices/contactus/nhscomplaints. Accessed May 3, 2016.
- 47. NI Direct . Make a complaint against the health service. 2016. https://www.nidirect.gov.uk/articles/make-complaint-against-health-service. Accessed May 3, 2016.
- 48. NHS England . What is PALS? 2016. http://www.nhs.uk/chq/Pages/1082.aspx. Accessed April 24, 2016.
- 49. The Scottish Association of Citizens Advice Bureaux . Comments and complaints. 2016http://www.cas.org.uk/pass. April 24, 2016.
- 50. Patient and Client Council NI . Making a complaint. 2016http://www.patientclientcouncil.hscni.net. Accessed April 24, 2016.
- 51. NHS England . NHS Choices. 2016. https://www.nhs.uk/pages/home.aspx. Accessed May 3, 2016.
- 52. iWGC . iWantGreatCare. 2016. https://www.iwantgreatcare.org. Accessed May 5, 2016.
- 53. Mumsnet Ltd . Mumsnet by parents for parents. 2016. https://www.mumsnet.com. Accessed May 5, 2016.
- 54. Patient Opinion . Patient Opinion Every Voice Matters. 2015. https://www.patientopinion.org.uk. Accessed May 5, 2016.
- 55. Baron S. Evaluating the patient journey approach to ensure health care is centred on patients. Nursing Times. 2009;105(22):20‐23. [PubMed] [Google Scholar]
- 56. Institute for Innovation and Improvement . Kinda Magic. Peninsula Community Health Community Interest Company; 2013.
- 57. Locock L, Robert G, Boaz A, et al. Testing accelerated experience‐based co‐design: a qualitative study of using a national archive of patient experience narrative interviews to promote rapid patient‐centred service improvement. Southampton UK: HMSO; 2014. [PubMed] [Google Scholar]
- 58. Royal College of Nursing . The 15 Steps Challenge Toolkit. 2014. https://www2.rcn.org.uk/__data/assets/pdf_file/0004/466078/01_The_15_Steps_Challenge_toolkit.pdf. Accessed July 12, 2016.
- 59. NHS Education for Scotland . Always Events. 2016. http://www.nes.scot.nhs.uk/education-and-training/by-theme-initiative/patient-safety-and-clinical-skills/always-events.aspx. Accessed February 23, 2017.
- 60. NHS England . Storytelling toolkit. 2015. https://www.england.nhs.uk/6cs/wp-content/uploads/sites/25/2015/09/scht-storytelling-toolkit.pdf. Accessed January 22, 2019.
- 61. NHS England . Friends and Family Test. 2014. https://www.nhs.uk/NHSEngland/AboutNHSservices/Documents/FFTGuide_Final_1807_FINAL.pdf. Accessed June 15, 2016.
- 62. Benson T, Potts HW. A short generic patient experience questionnaire: howRwe development and validation. BMC Health Serv Res. 2014;14:499. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 63. Reeves R, Seccombe I. Do patient surveys work? The influence of a national survey programme on local quality‐improvement initiatives. BMJ Quality & Safety. 2008;17:437‐441. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 64. Greaves F, Ramirez‐Cano D, Millett C, Darzi A, Donaldson L. Harnessing the cloud of patient experience: using social media to detect poor quality healthcare. BMJ Qual Saf. 2013;22(3):251‐255. [DOI] [PubMed] [Google Scholar]
- 65. Bate R, Robert G. Bringing user experience to healthcare improvement: the concepts, methods and practices of experience‐based design. Oxford, UK: Radcliffe Publishing; 2007. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.