Abstract
Purpose
The use of Patient-Reported Outcome Measures (PROMs) for individual patient management within clinical practice is becoming increasingly important. New evidence about graphic visualization formats for PROMs scores has become available. This systematic literature review evaluated evidence for graphic visualization formats of PROMs data in clinical practice for patients and clinicians, for both individual and group level PROMs data.
Methods
Studies published between 2000 and 2020 were extracted from CINAHL, PubMed, PsychInfo, and Medline. Studies included patients ≥ 18 years old in daily clinical practice. Papers not available in English, without full-text access, or that did not specifically describe visualization of PROMs data were excluded. Outcomes were: visualization preferences; interpretation accuracy; guidance for clinical interpretation.
Results
Twenty-five out of 789 papers were included for final analysis. Most frequently studied formats were: bar charts, line graphs, and pie charts. Patients preferred bar charts and line graphs as these were easy and quick for retrieving information about their PROMs scores over time. Clinicians’ interpretation accuracy and preferences were similar among graphic visualization formats. Scores were most often compared with patients’ own previous scores; to further guide clinical interpretation, scores were compared to norm population scores. Different ‘add-ons’ improved interpretability for patients and clinicians, e.g. using colors, descriptions of measurement scale directionality, descriptive labels, and brief definitions.
Conclusion
There was no predominant graphical visualization format approach in terms of preferences or interpretation accuracy for both patients and clinicians. Detailed clarification of graph content is essential.
Supplementary Information
The online version contains supplementary material available at 10.1186/s41687-022-00424-3.
Keywords: Patient reported outcome measures, Data visualization, Shared decision-making, Systematic review
Plain English summary
Patient-Reported Outcome Measures (PROMs) capture patients' self-reported health through the use of questionnaires. PROMs measure health related quality of life, daily functioning, and symptom experience, which are becoming increasingly important to incorporate in clinical practice for individual patient management. To present PROMs within clinical practice, raw or summarized PROMs scores can be visualized in graphical formats. To be useful during clinical encounters, both patients and clinicians ought to interpret such formats correctly. New evidence about graphic visualization formats for PROMs scores has become available. Therefore, we systematically reviewed the literature to evaluate evidence for graphic visualization formats of PROMs data in clinical practice. In 25 included papers, most studies used graphical formats like bar charts, line graphs, and pie charts for presenting PROMs scores. There was no predominant graphical visualization format approach in terms of preferences or interpretation accuracy for both patients and clinicians. Patients preferred bar charts and line graphs as these were easy and quick for retrieving information about their PROMs scores over time. Clinicians’ interpretation accuracy and preferences were similar among graphic visualization formats. The graphical interpretation of PROMs data for patients and clinicians can be improved by using colors, descriptions of measurement scale directionality, descriptive labels, and brief definitions.
Supplementary Information
The online version contains supplementary material available at 10.1186/s41687-022-00424-3.
Introduction
With an increasing emphasis on patient-centred care, there is a growing interest in outcome measures most relevant to patients [1–6]. Patient-reported outcomes measures (PROMs) comprise data collected from individual patients and include an array of outcomes such as symptoms, daily functioning, and health-related quality of life (HRQoL). PROMs are increasingly used in daily clinical practice for individual patient management [7]. Individual PROMs data incorporates the patient’s perspective on their health status and can detect issues that are most bothersome to the individual patient. By reporting these issues to both patients and clinicians, patient-physician communication improves [8–10]. This may support shared-decision making, and therefore offers considerable potential to enhance quality of care and clinical outcomes [8, 11]. A second application of PROMs data feedback is the use of aggregated PROMs scores collected in clinical studies or trials to inform patients about treatment harms and benefits [12].
Currently, when PROMs are incorporated within clinical practice, raw or summarized PROMs data are given as feedback to patients and/or clinicians by using different graphic visualization formats [13]. In order for them to understand and apply the information during clinical encounters, patients and clinicians ought to interpret such formats correctly [13]. A previous review by Bantug et al. [13] reported that the majority of patients and clinicians were able to interpret plain or straightforward graphs. Bantug et al. suggested that future research should focus on optimizing graphic visualization strategies. After the publication of this review in 2016, considerable new evidence has become available about this topic. Moreover, the focus of recent studies has shifted towards the effect of aspects such as score directionality [14–17] and axis labelling [1, 15] on correct interpretation by patients and clinicians. Furthermore, there is increased attention for guiding the clinical interpretation of PROMs data, e.g. to distinguish severe or clinically relevant symptoms [1, 3–5, 14, 18]. For instance, the display of score thresholds and warnings if scores change over time would be helpful in daily practice [1], as well as scores from reference populations to compare individual scores to [1, 5, 14]. Both facilitate the correct use of scores during clinical encounters. The focus on clinical interpretation led to the introduction of funnel plots [2, 19], heat maps [4], and icon arrays [3, 15, 18, 20], underlining the relevance of assessing a wider variety of graphic visualization formats.
While worldwide implementation of PROMs data collection in clinical practice keeps progressing [21], new evidence on graphic visualization formats for PROMs scores for interpretation by patients and clinicians has become available. This systematic literature review aims to (1) address the latest evidence for graphic visualization formats of PROMs in clinical practice, by extracting preferences and interpretation accuracy for patient and clinicians, and (2) investigate how clinically relevant PROMs scores are distinguished, in order to guide clinical interpretation of PROMs scores for their use during clinical encounters.
Methods
A systematic literature review was conducted according to the Preferred Reporting Items for Systematic Reviews and Meta-Analysis (PRISMA) guidelines [22].
Search strategy
An exhaustive search strategy was developed by a medical librarian (E.W.) experienced in systematic literature searches. The following search terms were included: (“Patient reported outcome (PRO)” OR “Health-Related Quality of Life”) AND “data presentation/display” AND “health communication” AND (“cancer” OR “clinical decision-making”). We focused the search on oncology as we work in an oncological setting. Also, most of these studies have been conducted in oncology, as confirmed by our search and previously shown by Bantug et al. [13]. Literature from January 2000 to July 2020 was searched in MEDLINE (accessed through PubMed), Embase (assessed through Ovid Platform), PsycINFO (assessed through Ovid Platform), and CINAHL. The full search strategy is included in Additional file 1: Table 1. Duplicates were removed using the Bramer method [23]. A forward and backward reference check was performed on all final included articles.
Review procedure
After removing duplicates, two researchers (E.A., I.F.) independently reviewed potential abstracts. The researchers eliminated articles according to the predefined inclusion and exclusion criteria (see below). In case of discrepancies, a third researcher (K.d.L.) was consulted. Subsequently, both researchers (E.A., I.F.) independently reviewed the full text version of each paper for inclusion in the final selection. Disagreements were discussed between the three researchers (E.A., I.F., and K.d.L.) to reach consensus.
Studies were included when concerning patients of 18 years and older treated in clinical practice; addressed communication of individual level or group level PROMs data, using graphic visualisation formats, to either patients or clinicians; were published in English. Exclusion criteria were: studies without English or full-text version; systematic literature reviews; Delphi studies; studies where PROMs had not been completed by patients or were not applied or visually presented to patients and/or clinicians.
Analyses
Two researchers (E.A., K.d.L.) independently extracted data from the included articles. The findings of both researchers were compared and verified by a third researcher (I.F.). First, study characteristics were described, including study goal, population, and type of PROMs data that was visualised. Second, findings about visualisation preferences and interpretation accuracy were extracted. Furthermore, methods and strategies for guiding the interpretation of scores during clinical encounters were extracted. The findings were described for patients and clinicians separately, and a distinction was made between individual PROMs data for use during clinical encounters, and mean or aggregated PROMs data that could be included in treatment decision-making.
Quality assessment
Two researchers (E.A., K.d.L.) independently assessed the methodological quality of the papers and compared their final judgments. The Critical Appraisal Skill Program (CASP) was used for methodological assessment of the included papers. CASP enables to systematically assess the trustworthiness, relevance, and results of published papers by comprising several criteria for qualitative studies, randomized controlled trials, and cohort studies [24]. The reviewers scored the papers per criteria with a positive or negative response, or ‘not applicable/unknown’. Studies with a positive score for half or more of the criteria were deemed of sufficient quality [24].
Results
Our search retrieved 1673 studies, from which 414 duplicates were excluded. Subsequently, from the remaining 1259 studies, 1186 were excluded; these did not describe PROMs data visualization (Fig. 1). Then, full-text articles from 73 eligible studies were assessed. From these, 47 were excluded based on inclusion and exclusion criteria, and/or because no full-text version was available (n = 15, all conference abstracts). Ultimately, 25 studies were included in this review for data extraction (Additional file 1: Table 2). After quality assessment following the CASP criteria, all studies had a positive score for half or more of the criteria (Additional file 1: Table 3).
Fig. 1.
PRISMA flow chart of literature search and review process
Most studies used either mixed methods design (n = 12), including human-centered design, or a qualitative design (n = 9), including interviews (Fig. 2). Sample sizes ranged from 8 (interview study) to 1,017 (survey study). Studies had been carried out in different clinical domains and in different countries, studying different graphic visualization formats and designs. The majority of studies included participants during or after treatment, whereas nine studies made use of hypothetical settings. PROMs data formats were either based on individual patient scores presented to patients (n = 17) and/or clinicians (n = 14), or based on mean group-reported data from for instance clinical trials that were presented to patients (n = 10) and clinicians (n = 8). The different graphic visualization formats that were studied are presented in Fig. 3. The results are presented according to the distinction between patients and clinicians, subdivided into preferences and interpretation accuracy.
Fig. 2.
Studies included in this review were carried out using a different designs; b in different countries; c investigating different graphic visualization formats*, and d among different clinical domains. RCT: Randomized Controlled Trial; USA: United States of America; UK: United Kingdom. *Studies included multiple formats
Fig. 3.
Different graphic visualization formats, presenting the domain of ‘emotional functioning’ as an example. Longitudinal: a line graph, including threshold line; b bar chart, including threshold line; c heat map. Cross-sectional: d icon array; e funnel plot; f pie chart
Individual level PROMs data—patients
An overview of the extracted data for patients on individual and group level is presented in Table 1.
Table 1.
Summary of data extraction: visualization strategies and preferences, interpretation accuracy, comparators; use of PRO data on individual and group level, in patients
Author | Primary study goal | Study population | Presenting PROMs data | Visualization | ||||
---|---|---|---|---|---|---|---|---|
Type of PROMs | What is presented? | Graphic visualisation format | Comparator | Outcomes of included studies preferences | Interpretation accuracy | |||
Individual level PROMs data visualization, patient | ||||||||
Brundage [16] | To investigate the interpretability of current PRO data presentation formats | N = 50 patients with variety of cancer diagnoses; N = 20 clinicians in active practice, Johns Hopkins Clinical Research Network (JHCRN)* | EORTC-QLQ-C30 scores | Individual scores and group means | Line graphs of mean scores | Previous scores | Simple line graphs for overall ease-of-understanding and usefulness | Patients accuracy ranged from 64–96% (line graphs questions) |
Tabulated scores | 92% preferred formats displaying multiple time points | A graph trending down with better = higher scores was correctly interpreted by 96%. A graph trending down up with better = lower scores was correctly interpreted by 64% | ||||||
Bubble plots of scores at a point in time | ||||||||
Heat map | ||||||||
Damman [14] | To investigate: | Interviews: patients with Parkinson's disease (N = 13) and clinicians (N = 14) | Not specified | Individual scores | Line graph | Patients with the same age, gender and disease duration | Bar chart is preferred (57.2%) compared to line graphs (42.3%) | What hindered easy comprehension: the use of a “higher = worse” directionality and comparative information of patients that are similar in terms of age, gender and disease progression |
(a) How patients and clinicians think about using PROMs during consultations; | Survey: patients (N = 115), the Netherland | Bar graph | Line and bar charts were interpreted most often correctly, compared with more “evaluative” formats like smileys and colors | |||||
(b) For which purpose patients and clinicians use PROMs during consultations; | Line graph with comparative data over time (i.e. average scores of similar patients) | Individual PROMs scores over time were interpreted more often correctly when presented in a bar chart (87.8%) compared to a line graph (74.3%) | ||||||
(c) How patients interpret PROMs information presented in various formats | ||||||||
Fischer [1] | To develop a PRO feedback report for mobile devices that is comprehensible and provides valuable information for patients after knee arthroplasty | Orthopedic patients (N = 8), Germany | Multiple (literature) | Individual scores | Text-based report and a graphical display (line graph, where scores are plotted over time, over a rainbow-colored background from red (bottom) to green (top) to visualize the grading of the individual scores) | Norm population | Short and condensed information using simple language (literature) | A text-based report is the least preferred but less susceptible to misinterpretation (literature) |
PROMIS (development) | An efficient way to present longitudinal PRO scores: graphs such as bar or line graphs (literature) | All participants correctly understood the line graph and were able to interpret the scores. Some needed some initial guidance on how to read a line graph | ||||||
Those (n = 3) in favor of graphs: easy and quick to get the relevant information from the line graph | The rainbow-colored background was understood by all participants | |||||||
The text-based (n = 2) version is easier to understand and most people are used to read short text messages | ||||||||
Geerards [26] | To assess the impact of tailored multimodal feedback and computerized adapted testing (CAT) on user experience in HRQoL assessment using validated PROMs | N = 1386 participants from the general population, United Kingdom (UK) | World Health Organization Quality of Life assessment (WHOQOL) | Individual scores | Graphical only | N/A | Respondents thought the questionnaire with graphical and text-based feedback was more interesting compared with no feedback assessment, whereas providing only graphical feedback did not make the questionnaire more interesting | 82.4% of patients thought the graphical feedback was accurate |
Graphical and adaptive text-based feedback | 92.9% of patients thought the graphical feedback was clear | |||||||
Graphs: Separate horizontal bar charts for 4 domains | ||||||||
Text: What each domain reflects, how score corresponds to average scores, and what score might mean | ||||||||
Grossman [27] | To identify the design requirements for an interface that assists patients with PRO survey completion and interpretation; to build and evaluate the interface of PROMs feedback | Interview: N = 13 patients with heart failure and N = 11 clinicians, study location or country was not described | Health IT Usability Evaluation Scale (Health-ITUES) | Individual scores | Small cards: Short sentence describing a severe symptom, which when clicked on provides textual educational information | N/A | Perceiving the mockup as useful and easy-to-use | Half of the participants failed to interpret the bar chart correctly, and even participants who could read it often required multiple attempts |
Usability testing: N = 12 patients with heart failure | Large cards: | Patients preferred visualizations with brief text descriptions | ||||||
Symptom name and description, visual representation of its severity, and a link to textual educational information | ||||||||
Graph: Bar chart (lists patient’s symptoms from most to least severe, with symptom’s severity scores) | ||||||||
Hartzler [6] | To conduct a HCD to engage patients, providers, and interaction design experts in the development of visual “PRO dashboards” that illustrate personalized trends in patients’ HRQoL following prostate cancer treatment | Focus groups (N = 60 patients) | Not specified | Individual scores | Pictographs | The dashboard compares patients’ trends with trends from “men like you” matched by default by age and treatment derived from a previously published prostate cancer cohort | Pictographs less helpful than bar charts, line graphs, or tables (P < .001) | Pictographs might reach patients with limited literacy |
N = 50 prostate cancer patients and N = 50 clinicians, study location or country was not described | Bar charts | Bar charts and line graphs are most preferred | Some patients expressed concern over inclusion of comparison scores without representation of data variability (e.g., confidence intervals, error bars), while others preferred simpler charts and graphs | |||||
Line graphs | ||||||||
Hildon [2] | To explore patients’ views of different formats and content of data displays of PROMs | N = 45 patients undergone or planning knee surgery in six focus groups, UK | Oxford Hip Score (OHS) | Individual scores | Different formats (table, bar chart, caterpillar and funnel plot) | N/A | Numerical tables lacked visual clarity | Representations of uncertainty were mostly new to the audience (numbers facilitated interpretation of uncertainty) |
Content (uncertainty displays, volume of outcomes, color, icons, and ordering) | Bar charts were liked because they were considered visually clear and facilitated appraisal at a glance, since it was a known format. But they do not give enough information | Traffic light colors were described as universally recognized | ||||||
Caterpillar plots were seen as visually clearer and to give more information but you would need to learn how to read them | Using colors consistently was important, as this enabled understanding across formats | |||||||
Funnel plots were difficult to read, had to learn how to read them | Stars were described as universally recognized and their interpretation did not require the ability to read | |||||||
Tables with icons were seen as accessible to the average person | The use of red and amber was thought to cause undue alarm while icons based on thumbs was seen as trivializing the issue | |||||||
Words (these were ‘at average’ ‘better’, ‘worse’, etc.) were liked because they were perceived as needing no personal interpretation | ||||||||
Izard [3] | To develop graphic dashboards of questionnaire responses from patients with prostate cancer to facilitate clinical integration of HRQoL measurement | N = 50 prostate cancer patients and N = 50 providers, USA | Expanded Prostate Cancer Index | Individual scores | Bar chart | Previous scores; ‘patients like me’ | 44% ranked bar chart dashboards as most preferred vs line graphs vs tables and pictographs | High reading scores for the table format |
Line graph | 20% found pictograph too complicated (too many steps to interpret) | |||||||
Table that display HRQOL data in raw form | 18% had difficulty disentangling facial expressions. Felt to be ‘‘too similar’’ | |||||||
Facial expression pictograph | 16% felt table to be easy to understand, 18% felt this format made HRQoL trends difficult to interpret | |||||||
Kuijpers [4] | To investigate patients’ and clinicians’ understanding of and preferences for different graphical presentation styles for individual-level EORTC QLQC30 scores | N = 548 cancer patients in four European countries and N = 227 clinicians, the Netherlands | EORTC QLQ-C30 | Individual scores | Bar chart with color | The preferred comparison group was one’s own previous results (40.9%) | 39% preferred colored bar charts, over heat maps (20%) and colored line graphs (12%) | Objective understanding did not differ between graphical formats |
Bar chart without color | ||||||||
Line graph with color | ||||||||
Line graph without color | ||||||||
Heat map | ||||||||
Liu [28] | To develop Rheumatoid Arthritis (RA) ‘dashboard’ that could facilitate conversations about PROs and is acceptable to a wide range of patients, including English and Spanish speakers, with adequate or limited health literacy | N = 25 RA patients and N = 11 clinicians from two academic rheumatology clinics, California | (1) Clinical Disease Activity Index (CDAI) | Individual scores | Line graph | Previous scores | Preference for more detailed information and more complex design in the adequate health literacy groups, but this preference was expressed by some limited health literacy participants as well | Several, particularly in the limited health literacy groups, did not notice or understand the longitudinal nature of data from left to right nor the temporal connection between the different graphic elements |
(2) Patient-Reported Outcomes Measurement Information System (PROMIS)-physical function scale | A few patients misinterpreted the line drawn between two data points to mean information from between the visits | |||||||
(3) Pain score | ||||||||
Loth [28] | To investigate patients’ understanding of graphical presentations of longitudinal EORTC QLQ-C30 scores | N = 40 brain tumor patients, Austria | EORTC QLQ-C30 | Individual scores | Colored bar chart | Previous scores | N/A | Objective correct answers about overall change was between 74.4% (fatigue) and 90.0% (emotional functioning) |
x | Thresholds based on reference population | Difficulties with correct interpretation of different directionality of the symptom and functioning scales | ||||||
Values below/above a predefined threshold for clinical importance were given as green (clinically unimportant) or red (clinically important) bars. Thresholds for clinical importance were distribution-based | The meaning of color-coding to highlight clinically important problems was answered correctly by 100% of patients (physical function and pain), and 92.5% (emotional function and fatigue) | |||||||
90% of the patients reported that the graphs (overall change) were “very easy” or “rather easy” to understand (subjective understanding) | ||||||||
Oerlemans [5] | To investigate whether patients with lymphoma wished to receive PRO feedback, including the option to compare their scores with those of their peers, and how this feedback was evaluated | Lymphoma patients (N = 64), the Netherlands | EORTC-QLQ-C30 + item tingling in hands or feet | Individual scores | Bar chart | Previous scores | Respondents had a slight preference for bar charts | 1 patient had trouble understanding the colors of the PRO feedback at first, but after looking for a second time it became clear |
Hospital Anxiety and Depression Scale (HADS) | Line graph | Reference population: | Preferred dotted line over a solid line to indicate “your score” in the bar chart | |||||
Adapted Self-Administered Comorbidity Questionnaire | General population | |||||||
Scores other lymphoma patients | ||||||||
Patients: The vast majority (94%) compared their scores with those of the lymphoma reference cohort and 64% compared their scores with those of the normative population without cancer, whereas 6% viewed only their own scores | ||||||||
Ragouzeos [25] | To develop a “dashboard” for RA patients to display relevant PRO measures for discussion during a routine RA clinical visit | Patients with rheumatology (N = 45) and providers (N = 12), USA | Not specified | Individual scores | Prototype PRO dashboard (on paper) | N/A | Important to show progress over time | Adding simple iconography and brief definitions of terms to the design helped patients understand which information the measured represented |
A longitudinal line graph with coloring helped patients see their measures as a process instead of a moment in time | ||||||||
Smith [18] | To improve formats for presenting individual-level PRO data (for patient monitoring) and group-level PRO data (for reporting comparative clinical studies) | N = 40 clinicians in active practice and N = 39 patients diagnosed with cancer ≥ 6 months previously, not currently receiving chemotherapy/radiation or within 6 months of surgery, from JHCRN* | Not specified | Individual scores | Line graphs | Previous scores | N/A | Ease-of-understanding ratings were high for all formats, with median ranges from 9–10 |
Pie charts | ||||||||
Bar charts | ||||||||
Icon array | ||||||||
Snyder [34] | To test approaches for presenting PRO data to improve interpretability | N = 627 cancer patients/survivors, N = 236 oncology clinicians, and N = 250 PRO researchers for survey, from JHCRN* | Not specified | Individual scores | 3 line-graphs: | Previous scores | N/A | 82–99% correctly responded across directionality items |
N = 10 patients and N = 10 clinicians for interviews | (1) Green-shaded normal range | 74–83% correctly identified domains that changed > 10 points | ||||||
(2) Red-circled possibly concerning scores | 53–86% accurately identified possibly concerning scores | |||||||
(3) Red threshold-lines between normal and concerning scores | Red circles were interpreted more accurately than green shading | |||||||
Higher = better were interpreted more accurately versus higher = more | ||||||||
Threshold-line significantly more likely to be rated “very” clear and most useful compared with green shading and red circles | ||||||||
Group level/aggregated PROMs data visualization, patients | ||||||||
Brundage [12] | To explore patients' attitudes toward, and preferences for, 10 visual and written formats for communicating Health Related Quality of Life (HRQoL) information | N = 14 men and N = 19 women with variety of cancer diagnoses, post treatment ≥ 6 months earlier, Canada | PRO results from hypothetical clinical trial (cross-sectional, longitudinal) | Group mean scores | Mean HRQL scores: | Two treatments | Line graphs were preferred, because of their relative simplicity and straightforward layout | N/A |
Trends in text | Decrease in preferences for line graphs when error bars around the mean are presented | |||||||
Mean scores | ||||||||
Mean scores with SD | ||||||||
Text | ||||||||
Change mean > 6 months | ||||||||
Brundage [30] | To determine which formats for presenting HRQoL data are interpreted most accurately and are most preferred by patients | Patients with variety of cancer diagnosis, previously treated (N = 198), Canada | PRO results from hypothetical clinical trial (cross-sectional, longitudinal) | Group mean scores | Two treatments | N/A | Line graphs were preferred, due to high ease of interpretation and perceived helpfulness | Line graphs most often interpreted correctly (98%), most easy to understand, and most helpful (all p < 0.0001) |
Format type, participant age and education independent predictors of accuracy rates | ||||||||
Brundage [16] | To investigate the interpretability of current PRO data presentation formats | N = 50 patients with variety of cancer diagnoses; N = 20 clinicians in active practice, from JHCRN* | EORTC-QLQ-C30 scores | Individual scores and group means | Line graph means over time | Two treatments | Simple line graphs were preferred, since they have a high ease-of-understanding and usefulness | Accuracy ranged from 36% (cumulative distribution function question) to 100% (line graph with confidence intervals question) |
Line graph with norms | Line graphs are straightforward and clear | Patients tented to find normed scores, p-values and confidence intervals confusing | ||||||
Line graph with confidence intervals | 87% of patients preferred formats displaying multiple time-points | |||||||
Bar chart of average changes | ||||||||
Bar chart with | ||||||||
definition (improved, stable, worsened) | ||||||||
Cumulative distribution functions | ||||||||
Damman [14] | To investigate: | Interviews: patients with Parkinson's disease (N = 13) and clinicians (N = 14) | Not specified | Individual scores | Line graph with results of 2 treatment options | Patients with the same age, gender and disease duration | 56% of patients found line graphs most useful | Line graph showing results of two treatment options resulted in decisions reflecting adequate comprehension of information |
(a) How patients and clinicians think about using PROMs during consultations; | Survey: patients (N = 115), the Netherlands | Bar chart with results of 2 treatment options | 47% of patients found bar charts most useful | |||||
(b) For which purpose patients and clinicians use PROMs during consultations; | Bar chart with performance of 2 providers | 43% of patients found information with performance of two providers useful | ||||||
(c) How patients interpret PROMs information presented in various formats | ||||||||
McNair [32] | To assess patients’ understanding of multidimensional PROs in a graphical format | Patients with esophageal and gastric cancer (N = 132), UK | Semi-structured interviews | Mean scores | Line graphs: | Two treatments | N/A | 87% of patients accurately interpreted multidimensional graphical PROs from two treatments |
(1) Treatment changes in a single PRO over time | 81% of patients was able to interpret graph 4 correctly | |||||||
(2) Different PRO, reversed direction of treatment | 67% of patients was able to integrate information from two graphs together | |||||||
(3) Divergent and convergent PROs | ||||||||
(4) Divergent and convergent PROs over 18 months | ||||||||
Smith [18] | To improve formats for presenting individual-level PRO data (for patient monitoring) and group-level PRO data (for reporting comparative clinical studies) | N = 40 clinicians in active practice and N = 39 patients diagnosed with cancer ≥ 6 months previously, not currently receiving chemotherapy/radiation or within 6 months of surgery, from JHCRN* | Not specified | Individual scores, proportional data | Line graphs | Previous scores | 55% of patients preferred pie charts | N/A |
Pie charts | 25% of patients preferred icon arrays | |||||||
Bar charts | 20% of patients preferred bar charts | |||||||
Icon array | 45% of patients preferred formats with an asterisk indicating important differences | |||||||
Tolbert [29] | To identify the association of PRO score directionality and score norming on a) how accurately PRO scores are interpreted and b) how clearly they are rated by patients, clinicians, and PRO researchers | N = 629 patients (various oncologic diagnoses, N = 139 oncology clinicians, and N = 249 PRO researchers, conducted at the Johns Hopkins Clinical Research Network (JHCRN)* | Two treatments | Mean scores | Line graph 3 versions: | Two treatments | 84% of patients rated “Better” formatted line graphs most often as “very clear” or “somewhat clear” | 56% of patients answered questions correctly for “better” formatted lines, compared to 41% for “more” and 39% for “normed” graphs |
(1) Lines going up indicating “better” outcomes | The normed value confused patients | |||||||
(2) Lines going up indicating “more” (better for function domains, worse for symptoms). (3) Lines “normed” to a general population | ||||||||
Tolbert [20] | To identify best practices for presenting PRO results expressed as proportions of patients with changes from baseline (improved/ stable/ worsened) for use in patient educational materials and decision aids | N = 629 patients (various oncologic diagnoses, treated), N = 139 oncology clinicians, and N = 249 PRO researchers, conducted at the Johns Hopkins Clinical Research Network (JHCRN)* | Two treatments | Proportions | Pie chart | Two treatments | Preferred pie charts: these were easy to read and enabled obtaining information quickly. Rated the clearest for communicating proportions changed from baseline | Patient’s accuracy was highest for pie charts and icon arrays |
Bar chart | Noted helpful aspects of bar charts: “Side by side comparisons are much easier to read and comprehend” | Bar graphs were less accurately interpreted than pie charts and icon arrays | ||||||
Icon array | Icon arrays would be easy to understand for patients | |||||||
Van Overveld [19] | To investigate the preferences of receiving feedback between stakeholders | N = 37 patients, medical specialists, allied health professionals and health insurers in the Netherlands | Audit data on professional practice and health care outcomes | National average scores | Bar graph | National average scores | Patients preferred both a pie chart and a bar chart | Give feedback with average national scores |
Pie chart | Patients prefer a figure over plain text | National average scores on indicators of more interest for patient organizations and professionals | ||||||
Line graph | ||||||||
Point graph | ||||||||
Area graph | ||||||||
Box plot | ||||||||
Kaplan- Meier graph | ||||||||
Funnel plot |
Definitions: individual level PROMs data—The patient’s perspective on their health status; Group level PROMs data—Aggregated PROMs scores collected in clinical studies or trials
CDAI—Clinical Disease Activity Index; EHR—Electronic Health Record; EORTC-QLQ-C30—European Organization for Research and Treatment of Cancer Quality of life questionnaire Core 30; HADS—Hospital Anxiety and Depression Scale; HCD—Human Centered Design; HRQoL—Health-Related Quality of Life; HUI—Health Utility Index; Health-ITUES—Health IT Usability Evaluation Scale; JHCRN—Johns Hopkins Clinical Research Network; N/A—Not Applicable; PRO(s) —Patient Reported Outcome(s); PROMs—Patient Reported Outcome Measurements; PROMIS—Patient-Reported Outcomes Measurement Information System; QoL—Quality of Life; REALM-SF—Rapid Estimate of Adult Literacy in Medicine Short Form; SD—Standard Deviation; WHOQOL—World Health Organization Quality of Life
*JHCRN—Johns Hopkins Clinical Research Network: A consortium of academic and community health systems in the US mid-Atlantic with clinics outside the USA as well
Preferences
From the 25 studies included in this study, 12 assessed patients’ preferences. Patients appreciated that their individual PROMs scores were presented visually over time [16, 25]. Six studies reported that patients preferred bar charts over other graphic visualization formats for reporting longitudinal PROMs data [1, 3–6, 14]. Bar charts were ‘visually clear’, and facilitated ‘appraisal at a glance’, since bar charts are a generally well-known format for presenting data [2]. Five studies reported that line graphs were preferred as well [1, 6, 15, 16, 25]. Patients experienced line graphs as ‘easy to understand’ and ‘quick to retrieve relevant information from’ regarding their symptoms and well-being [1]. Furthermore, line graphs were preferred for overall ease-of-understanding and usefulness [16].
Interpretation accuracy
Eight studies assessed patients’ interpretation accuracy for visualizing individual PROMs data. Bar charts were more often interpreted correctly (87.8%) compared to line graphs (74.3%) [14]. According to Geerards et al. [26], a study conducted among 1386 participants of the general population, 92.9% of participants thought that bar charts were clear, and 82.4% thought bar charts were accurate. In a study by Loth et al. [17], among 40 patients, 90% reported that bar charts were “very easy” or “rather easy” to understand. Nevertheless, Grossman et al. [27] described that from 12 included patients, half of them failed to interpret bar charts correctly, and even patients who successfully read it, often required multiple attempts to do so. These participants were, however, introduced to bar charts for the first time [27]. One study reported that line graphs could be interpreted accurately by patients; nonetheless, an undefined proportion of patients needed initial guidance on how to read line graphs [1]. Liu et al. [28] reported that from 25 respondents (patients), several did not notice or understand the longitudinal nature of a line graph from left to right, particularly those with limited health literacy. Furthermore, a few patients misinterpreted a line drawn between two data points to mean information from between visits [28].
Group-level or aggregated PROMs data—patients
The aim of presenting aggregated or mean PROMs scores to patients is to inform them about potential treatment outcomes, in order to support shared decision-making. In the included studies, aggregated PROMs data were mainly derived from clinical studies or RCTs. These were often presented to patients to (hypothetically) inform them about differences in outcomes between two treatments [12, 15, 20, 29–33].
Preferences
Seven studies assessed patients’ preferences for presenting aggregated outcomes over time. Four studies reported that patients preferred line graphs over bar and pie charts [12, 14, 16, 30]. Line graphs were preferred because of their relative simplicity, straightforward layout, and perceived helpfulness [12, 16, 30]. The preference for line graphs decreased when confidence intervals were presented [12]. To present proportional data, three studies reported that patients preferred pie charts [18–20]. Pie charts were easy to read and enabled obtaining information quickly [20]. Smith et al. [18] reported that of 39 patients, 55% preferred pie charts, compared to a 25% preference for icon arrays, and a 20% preference for bar charts. Van Overveld et al. [19] reported that among 37 patients, pie and bar charts were both preferred.
Interpretation accuracy
Seven studies assessed patients’ interpretation accuracy for presenting aggregated outcomes over time. In four studies, patients’ most often correctly interpreted line graphs [14, 16, 29, 30]. Interpretation accuracy ranged from 98–100% [16, 30]. In particular, 56% of patients correctly answered questions for “better” (i.e. higher score means better outcomes) formatted lines, compared to 41% for “more” (i.e. higher score means more symptoms), and 39% for “normed” (i.e. score as compared to reference scores) line graphs [29]. Line graphs showing results of two treatment options resulted in decisions that reflected adequate comprehension of the information in the graph [14]. One study reported that patients’ accuracy was highest for pie charts and icon arrays, compared to bar charts [20]. Icon arrays were easy to understand for patients as well [20].
Individual level PROMs data—clinicians
An overview of the extracted data for clinicians on individual and group level is presented in Table 2.
Table 2.
Summary of data extraction: visualization strategies and preferences, interpretation accuracy, comparators; use of PRO data on individual and group level, in clinicians
Author | Primary study goal | Study population | Presenting PROMs data | Visualization | Outcomes of included studies | |||
---|---|---|---|---|---|---|---|---|
Type of PROMs | What is being presented? | Graphic visualisation format | Comparator | Preferences | Interpretation accuracy | |||
Individual level PROMs data visualization, clinicians | ||||||||
Brundage [16] | To investigate the interpretability of current PRO data presentation formats | N = 50 patients with variety of cancer diagnoses; N = 20 clinicians in active practice, from Johns Hopkins Clinical Research Network (JHCRN)* | EORTC-QLQ-C30 scores | Individual scores and group means | Two line graphs: | Previous scores | Preference for line graphs: overall ease-of-understanding and usefulness | Interpretation accuracy was high across formats |
Higher = better functioning | 90% of clinicians preferred formats displaying multiple time-points vs single time-points | Inconsistency between whether higher scores were better or worse contributes to incorrect accuracy (Uptrend lines intuitively signify improvement of understanding) | ||||||
Higher = more symptoms | Tabulated scores were considered boring but straightforward | |||||||
Bubble plot | ||||||||
Heat maps | ||||||||
Brundage [37] | To evaluate the interpretation accuracy and perceived clarity of various strategies for displaying clinical trial PRO findings | Oncology clinicians (N = 233) and PRO researchers (N = 248), from JHCRN* | PRO results from hypothetical clinical trial (cross-sectional, longitudinal) | Longitudinal individual data, proportions | Bar chart | Line graph also normed against general population | Not one approach for either longitudinal data or proportions changed is universally appealing, nor is free of misinterpretation errors | Line graph: |
Pie chart | More likely to be interpreted correctly “better” vs “normed” graphs (p = 0.04) | |||||||
3 Line graphs: | No differences between “better” and “more” | |||||||
1) Higher = “better” functioning; | Regardless of graph type and version, adding asterisks for clinical significance and confidence limits did not contribute to better interpretation accuracy | |||||||
2) Higher = “more” symptoms; | Bar chart/pie chart: | |||||||
3) “Normed” against general population | Respondents less likely to make interpretation errors with pie vs bar charts (p < 0.001) | |||||||
Odds of selecting an “incorrect” treatment significantly lower in pie charts compared to bar charts | ||||||||
Clarity ratings did not differ between formats | ||||||||
Damman [14] | To investigate: | Interviews: patients with Parkinson's disease (N = 13) and clinicians (N = 14) | Not specified | Individual scores | Line graph | Patients with the same age, gender and disease duration | Strong preference for individual PROMs data over time | Individual PROMs scores with comparative data of similar patients were found useful, some expressed doubts |
(a) How patients and clinicians think about using PROMs during consultations; | Survey: patients (N = 115), the Netherlands | Bar graph | Line and bar graphs | |||||
(b) For which purpose patients and clinicians use PROMs during consultations; | Line graph with comparative data over time (i.e. average scores of similar patients) | Scores from repeated measurements over time | ||||||
(c) How patients interpret PROMs information presented in various formats | Multiple individual quality of life, rather than one overall quality of life score | |||||||
Identified the possibility to use aggregated PROMs scores as evidence for treatment options | ||||||||
Grossman [27] | To identify the design requirements for an interface that assists patients with PRO survey completion and interpretation; to build and evaluate the interface of PROMs feedback | Interview: N = 13 patients with heart failure and N = 11 clinicians, study location or country was not described | Health IT Usability Evaluation Scale (Health-ITUES) | Individual scores | Small cards: Contains a short sentence to describe a severe symptom, which when clicked on provides textual educational information | N/A | Perceiving the mockup as useful and easy-to-use | Two providers reported that PROs might reduce their cognitive load |
Usability testing: N = 12 patients with heart failure | Graph: | Preference for tracking symptoms over time | ||||||
Bar chart that lists the patient’s symptoms from most to least severe and displays each symptom’s severity score | ||||||||
Large cards: | ||||||||
Displays a symptom name, its two-sentence description, a visual representation of its severity, and a link to textual educational information | ||||||||
Hartzler [36] | To share lessons learned from engaging clinicians to inform design of visual dashboards | Clinicians: N = 12 for interviews, N = 40 for surveys and consensus meeting, N = 9 for user testing, study location or country was not described | Not specified | Individual scores | PRO data needs appear to differ for health care providers and administrative staff as key target users | N/A | Participants liked viewing trends over time | Value was found in developing meaningful ways to report on this new source of data |
The functional prototype has 3 components: | Participants found the views to provide a useful basis for comparison | In addition to the information buttons provided on “how to interpret this chart,” clear labels are needed, such as on chart axes | ||||||
(1) An “At a glance” screen providing a simple data overview of PROs data | ||||||||
(2) An “Analyze” screen providing a data view the user can filter | ||||||||
(3) A “Data quality” screen | ||||||||
Hartzler [6] | To conduct a HCD to engage patients, providers, and interaction design experts in the development of visual “PRO dashboards” that illustrate personalized trends in patients’ HRQoL following prostate cancer treatment | Focus groups (N = 60 patients) | Not specified | Individual scores | Pictographs | The dashboard compares patients’ trends with trends from “men like you” matched by default by age and treatment derived from a previously published prostate cancer cohort | Pictographs less helpful than bar charts, line graphs, or tables (P < .001) | Pictographs might reach patients with limited literacy |
N = 50 prostate cancer patients and N = 50 clinicians, study location or country was not described | Bar charts | Preferred bar charts and line graphs | Some participants, both patients and providers, expressed concern over inclusion of comparison scores without representation of data variability (e.g., confidence intervals, error bars), while others preferred simpler charts and graphs | |||||
Line graphs | ||||||||
Izard [3] | To develop graphic dashboards of questionnaire responses from patients with prostate cancer to facilitate clinical integration of HRQoL measurement | N = 50 prostate cancer patients and N = 50 providers from Seattle, USA | Expanded Prostate Cancer Index | Individual scores | Bar chart | Previous scores; ‘patients like me’ | No universally preferred dashboard format: 30% preferred tables, 34% preferred bar charts, and 34% preferred line graphs | Helpfulness and confidence ratings varied among dashboard format. Pictographs had the lowest helpfulness compared with table, bar, and line graph formats |
Line graph | ||||||||
Table that display HRQOL data in raw form | ||||||||
Facial expression pictograph | ||||||||
Jagsi [35] | To investigate practicing oncologists view on incorporating routine collection of PROs into cancer care | N = 17 oncologists, USA | Edmonton Symptom Assessment System | Individual scores | Bar chart | Previous scores | Ability to track symptoms over time and effect of intervention | Keep it simple: limit number of symptoms |
Link number scale to narrative | ||||||||
Kuijpers [4] | To investigate patients’ and clinicians’ understanding of and preferences for different graphical presentation styles for individual-level EORTC QLQC30 scores | N = 548 cancer patients in four European countries and N = 227 clinicians, the Netherlands | EORTC QLQ-C30 | Individual scores | Bar chart with color | The preferred comparison group was one’s own previous results (40.9%) | Medical specialist: | Medical specialist: |
Bar chart without color | Heat map (46%) | Objective understanding of 78% | ||||||
Line graph with color | Nurses: | Nurses: | ||||||
Line graph without color | Bar chart (non-colored) and heat map (32%) | Objective understanding of 74% | ||||||
Heat map | 85% of all HCP’s indicated that the graphs were (easy) to understand, not differing between professions or graphical presentation styles | |||||||
Uniformity in scoring direction would be preferred | ||||||||
A significant difference for overall change scores, with the non-colored bar charts being interpreted correctly more often than the other graphical displays | ||||||||
Ragouzeos [25] | To develop a “dashboard” for RA patients to display relevant PRO measures for discussion during a routine RA clinical visit | Patients with rheumatology (N = 45) and providers (N = 12), USA | Not specified | Individual scores | Prototype PRO dashboard (on paper) | Previous scores | N/A | Information needs to be clearly focused on what is most essential |
Important to show progress over time | ||||||||
Santana [33] | To describe the process, feasibility and acceptability of use of the Health Utilities Index (HUI) in routine clinical care | Pre- and post-heart and -lung transplant patients (N = 151), Canada | Health Utilities Index (HUI) | Individual scores | HUI score card, using a color-coded system | Pre- and post-treatment scores | N/A | Clinicians did not need much time to understand the use of the HUI score card |
Clinicians developed their own way of using the information over time | ||||||||
Smith [18] | To improve formats for presenting individual-level PRO data (for patient monitoring) and group-level PRO data (for reporting comparative clinical studies) | N = 40 clinicians in active practice and N = 39 patients diagnosed with cancer ≥ 6 months previously, not currently receiving chemotherapy/radiation or within 6 months of surgery, from JHCRN* | Not specified | Individual scores, proportional data | Line graphs | Previous scores | 75% preferred the line graph | Ease-of-understanding ratings were high for all formats |
Pie charts | Directional inconsistency emerged as an interpretation challenge | |||||||
Bar charts | Format interpretation challenges included explaining the meaning of scores (i.e., whether scores are good/bad, what normal is) | |||||||
Icon array | ||||||||
Snyder [34] | To test approaches for presenting PRO data to improve interpretability | N = 627 cancer patients/survivors, N = 236 oncology clinicians, and N = 250 PRO researchers for survey, from JHCRN* | Not specified | Individual scores | 3 line-graphs: | Previous scores | N/A | The proportion responding correctly across the 4 directionality items ranged from 80 to 100% |
N = 10 patients and N = 10 clinicians for interviews | 1) Green-shaded normal range | Red circles were interpreted more accurately than green shading | ||||||
2) Red-circled possibly concerning scores | Higher = better were interpreted more accurately versus higher = more | |||||||
3) Red threshold-lines between normal and concerning scores | Threshold-line significantly more likely to be rated “very” clear and most useful compared with green shading and red circles | |||||||
Group level/aggregated PROMs data visualization, clinicians | ||||||||
Brundage [16] | To investigate the interpretability of current PRO data presentation formats | N = 50 patients with variety of cancer diagnoses; N = 20 clinicians in active practice, from JHCRN* | EORTC-QLQ-C30 scores | Individual scores and group means | Line graph means over time | Two treatments (study arms) | 90% of clinicians preferred formats displaying multiple time-points | Line graphs contributed to overall ease-of-understanding and usefulness |
Line graph with norms | Preference for line graphs of normed sores or with confidence intervals | Normed scores provided basis for comparison beyond two treatments, p-values and confidence intervals were particularly important for publications | ||||||
Line graph with confidence intervals | Some preference for bar charts to compare treatments | Cumulative distributing function confusing and difficult to interpret | ||||||
Bar chart of average changes | Inconsistency between whether higher scores were better or worse contributes to incorrect accuracy | |||||||
Bar chart with definition (improved, stable, worsened) | ||||||||
•Cumulative distribution functions | ||||||||
Damman [14] | To investigate: | Interviews: patients with Parkinson's disease (N = 13) and clinicians (N = 14) | Not specified | Individual scores | Line graph with results of 2 treatment options | Patients with the same age, gender and disease duration | Identified the possibility to use aggregated PROMs scores as evidence for treatment options | N/A |
(a) How patients and clinicians think about using PROMs during consultations; | Survey: patients (N = 115), the Netherlands | Bar chart with results of 2 treatment options | Aggregated PROMs scores for provider options could be useful, but would not be used much in clinical practice | |||||
(b) For which purpose patients and clinicians use PROMs during consultations; | Bar chart with performance of 2 providers | |||||||
(c) How patients interpret PROMs information presented in various formats | ||||||||
Liu [28] | To develop Rheumatoid Arthritis (RA) ‘dashboard’ that could facilitate conversations about PROs and is acceptable to a wide range of patients, including English and Spanish speakers, with adequate or limited health literacy | N = 25 RA patients and N = 11 clinicians from two academic rheumatology clinics, California (USA) | (1) Clinical Disease Activity Index (CDAI) | Individual scores | Line graph | Aggregated clinical data | A dashboard is a potential method for aggregating data from various sources | A ‘snapshot’ of relevant information for a particular patient would make HCP’s own medical decisions easier |
(2) Patient-Reported Outcomes Measurement Information System (PROMIS)-physical function scale | Clinicians were very interested in customizing the dashboard to their own needs and recommended that it can be designed to present information that is more detailed | |||||||
(3) Pain score | ||||||||
Smith [18] | To improve formats for presenting individual-level PRO data (for patient monitoring) and group-level PRO data (for reporting comparative clinical studies) | N = 40 clinicians in active practice and N = 39 patients diagnosed with cancer ≥ 6 months previously, not currently receiving chemotherapy/radiation or within 6 months of surgery, from JHCRN* | Not specified | Individual scores, proportional data | Line graphs | Average changes | For proportional data formats: pie charts (70%) | Median ease-of-understanding ranged from 6.5 to 8 |
Pie charts | Few clinicians (10%) preferred bar charts | Mixed feelings about indications of clinical significance in terms of having p-values in addition to confidence intervals and asterisks indicating important differences | ||||||
Bar charts | 75% preferred the line graph | Directional inconsistency emerged as an interpretation challenge | ||||||
Icon array | Format interpretation challenges included explaining the meaning of scores (i.e., whether scores are good/bad, what normal is) | |||||||
Tolbert [29] | To identify the association of PRO score directionality and score norming on a) how accurately PRO scores are interpreted and b) how clearly they are rated by patients, clinicians, and PRO researchers | N = 629 patients (various oncologic diagnoses), N = 139 oncology clinicians, and N = 249 PRO researchers, conducted at the Johns Hopkins Clinical Research Network (JHCRN)* | Two treatments | Mean scores | Line graph 3 versions: | Two treatments | “Better” formatted line graphs were rated most often as “very clear” or “somewhat clear” by all three groups (84% by patients, 81% by clinicians, and 85% by researchers) | Answers correct for “better”: 70%; “more”: 65%; “normed”: 65% |
(1) Lines going up indicating “better” outcomes | However, the range in the proportion rating each format “very clear” or “somewhat clear” was narrow: 77% to 85% | “More” line graph comments noted that up could mean different things, which could lead to errors | ||||||
(2) Lines going up indicating “more” (better for function domains, worse for symptoms). 3) Lines “normed” to a general population average of 50 | “Better” line graph comments pointed out how changing the scale could result in interpretation errors as one must orient to the direction of the scale each time | |||||||
Tolbert [20] | To identify best practices for presenting PRO results expressed as proportions of patients with changes from baseline (improved/ stable/ worsened) for use in patient educational materials and decision aids | N = 629 patients (various oncologic diagnoses, treated), N = 139 oncology clinicians, and N = 249 PRO researchers, conducted at the Johns Hopkins Clinical Research Network (JHCRN)* | Two treatments | Proportions changed | Pie chart | Two treatments | Preferred pie charts: these were easy to read and enabled obtaining information quickly | Clinician and researchers scored pie charts as the most accurately interpreted |
Bar chart | 43% had positive feedback on icon arrays | In general, bar graphs were less accurately interpreted than pie charts and icon arrays | ||||||
Icon array | 38% had positive feedback on bar charts | Noted helpful aspects of bar charts: “Side | ||||||
by side comparisons are much easier to read and comprehend” | ||||||||
van Overveld [19] | To investigate the preferences of receiving feedback between stakeholders | N = 37 patients, medical specialists, allied health professionals and health insurers in the Netherlands | Audit data on professional practice and health care outcomes | National average scores | Bar graph | National average scores | Preference for bar charts since they are easier to read | Box plots, Kaplan–Meier graphs and funnel plots give a less clear overview and are more difficult to interpret |
Pie chart | For survival and process indicators: Kaplan–Meier graphs and box plots | Find a balance between giving feedback and giving too much information | ||||||
Line graph | Give an overview of the results first, followed by the details | |||||||
Point graph | Present it that one can easily understand without explanation | |||||||
Area graph | ||||||||
Box plot | ||||||||
Kaplan- Meier graph | ||||||||
Funnel plot |
Definitions: individual level PROMs data—The patient’s perspective on their health status; Group level PROMs data—Aggregated PROM scores collected in clinical studies or trials
Abbreviations: CDAI—Clinical Disease Activity Index; EHR—Electronic Health Record; EORTC-QLQ-C30—European Organization for Research and Treatment of Cancer Quality of life questionnaire Core 30; HADS—Hospital Anxiety and Depression Scale; HCD—Human Centered Design; HRQoL—Health-Related Quality of Life; HUI—Health Utility Index; Health-ITUES—Health IT Usability Evaluation Scale; JHCRN—Johns Hopkins Clinical Research Network; N/A—Not Applicable; PRO(s) —Patient Reported Outcome(s); PROMs—Patient Reported Outcome Measurements; PROMIS—Patient-Reported Outcomes Measurement Information System; QoL—Quality of Life; REALM-SF—Rapid Estimate of Adult Literacy in Medicine Short Form; SD—Standard Deviation; WHOQOL—World Health Organization Quality of Life
*JHCRN—Johns Hopkins Clinical Research Network: A consortium of academic and community health systems in the US mid-Atlantic with clinics outside the USA as well
Preferences
Thirteen studies assessed clinicians’ preferences regarding visualization of PROMs scores from individual patients. In general, clinicians appreciated viewing PROMs scores repeated over time, in order to track their patients’ symptom experiences [14–16, 25, 35, 36]. Moreover, six studies showed that bar charts were most preferred when plotting longitudinal individual PROMs data [1, 2, 4, 5, 14, 17]. Furthermore, line graphs were preferred in four studies [1, 14–16, 29]. However, Brundage et al. [37] and Izard et al. [3] both stated that clinicians did not universally find one approach for longitudinal data as appealing or preferred. Contrastingly, a study among 227 health professionals by Kuijpers et al. [4] showed that the majority of medical specialists (46%) and nurses (32%) preferred heat maps to line graphs and bar charts.
Interpretation accuracy
Thirteen studies assessed clinicians’ interpretation accuracy of graphic visualization formats for individual PROMs level data. Multiple studies showed that clinicians’ interpretation accuracy was similar over different graphic visualization formats: both line graphs and bar charts were found easy to understand, and were interpreted accurately [3, 4, 18]. Hartzler et al. [36] reported that among twelve clinicians, respondents generally felt that both line graphs and bar charts provide a useful basis for comparison. Pictographs were reported to be the least helpful for clinicians, although clinicians stated these could be helpful for patients with limited literacy [3, 6]. Uniformity in directionality of scores could increase clinicians’ interpretation accuracy in different graph formats. However, Brundage et al. [37] found that adding asterisks for clinical significance and confidence limits around scores did not contribute to a better interpretation accuracy in clinicians.
Group-level or aggregated PROMs data—clinicians.
Preferences
When presenting aggregated PROMs data, seven studies stated that clinicians mostly compared data between two treatments or compared scores to mean reference population scores. Brundage et al. [16] stated that for comparing treatments, 18 out of 20 clinicians preferred formats displaying multiple time-points, with the highest preference for line graphs with normed scores or confidence intervals, or bar charts. Smith et al. and Tolbert et al. [18, 29] describe a preference for line graphs, with over 75% of included clinicians preferring this graphical format; bar charts were less supported, as only 10% of 40 clinician participants preferred bar charts [18]. Van Overveld et al. [19] reported however that bar charts were preferred, since they were easy to read. To present proportional data, pie charts were most preferred in 3/8 studies [18, 20, 34]. To meet a variety of preferences, Liu et al. [28] found that a dynamic dashboard gave clinicians the opportunity to customize the formats to their own needs.
Interpretation accuracy
Six studies assessed clinicians’ interpretation accuracy on aggregated data. Line graphs usually contributed to the ease of understanding PROMs scores for clinicians [16, 29]. Additionally, pie charts with proportional data were most often interpreted accurately by clinicians, for example when pie charts presented a proportional change in outcomes compared to baseline [20]. Clinicians both endorsed and objected that p-values, confidence intervals, and normed scores could contribute to their interpretation accuracy of aggregated PROMs data [16, 18]. Furthermore, directional consistency, balancing information, and giving feedback, and clear labeling could improve interpretation accuracy [18, 19].
Throughout many of the included studies, challenges were described that may affect correctly interpreting visualized PROMs data, such as (1) Directional inconsistency, i.e. a higher functioning score means better health, but a lower symptom score means better health as well [5, 16, 17, 34, 37]; (2) Lack of standardisation rules for interpretation and visualization, that may cause interpretation inaccuracy [2, 25, 35]; (3) The need to designing multiple formats per target group, as no ‘one-size-fits-all’ solution in graphic visualisation exists for both patients and clinicians [3, 6, 28]; (4) The timing of providing feedback on PROMs visualization, as this affects assessment experience [14, 26], and (5) Patients ‘ opposition to PROMs use in clinical practice [19, 36]. We summarize the challenges and the proposed solutions for these challenges in Table 3.
Table 3.
Challenges and factors for improvement to consider when implementing visual individual PROMs feedback in clinical practice
Challenges that may hinder graphic visualization format interpretation | Possible factors to improve graphic visualization format interpretation |
---|---|
Patients and clinicians | |
Directional inconsistency in longitudinal data (i.e., sometimes higher scores can mean better or worse) | Make use of standard descriptive labels (consider using ‘better*’ instead of ‘normed**’ or ‘more***’ for describing directionality of scores) [34, 37] |
Preferred by 79% of patients and 90% of clinicians when concerning individual level PROMs data and 100% of clinicians when concerning group level PROMs data | |
Consistent use of clear ratings: higher scores are always better results (i.e. in some frequently used PROMs, higher score are better when scores describe functioning, but lower score are better when symptom burden is described. This causes interpretation challenges) [37] | |
Indicate with an arrow on the y-axis which direction means the score is better [16] | |
Describe directionality by plain text that is understandable despite literacy or education level [5] | |
Provide detailed information on the meaning of high and low score [17] | |
Interpretation accuracy of what exact PROMs information is represented in the graphic visualization format | Provide an instructive aid for patients and clinicians [2] |
Use simple iconography for demonstrating single PROMs values [25] | |
Use brief definitions of different PROMs domains/values [25] | |
Limit the number of presented symptoms per graphic visualization format [35] | |
No ‘one-size-fits-all’ solution | Make use of a dynamic dashboard, which can display multiple types of visualization strategies. Thereby, you provide users the ability to select a preferred format instead, including the ability to add or remove dashboard elements such as error bars and shading [3, 6, 28] |
Developing a clinic-based video tutorial for the dashboard to explain what is shown on the dashboard and how the patient and clinician can customize the dashboard to their needs [28] | |
Patients | |
Interpretation accuracy of what exact PROMs information is represented in the graphic visualization format | Ask patients to prioritize their symptoms, to avoid an overload of information [35] |
Timing of providing feedback on PROMs visualization | Provide feedback immediately after assessment, and before consultation, to significantly improve assessment experience when providing combined graphical and tailored text-based feedback [14, 26] |
Patients ‘ opposition to PROMs use in clinical practice | Ask permission to the patient to receive their own results and/or the results of the general population [19] |
Provide information so patients know what PROMs data might show and how their practice might change [36] | |
Tell patients that data is trustworthy and are handled confidentially [19, 36] | |
Do not provide anonymous feedback [19] | |
Visualize as transparently as possible what type of care is delivered [19] | |
Clinicians | |
Interpretation accuracy of what exact PROMs information is represented in the graphic visualization format | Eliminate comparison groups or inform comparison group scores with confidence intervals or error bars [3], to better counsel the patients about their score (makes it easier to understand) |
Link the PROMs outcome scores (scale in the graphic visualization format) to the meaning of the narrative (i.e.; tell the patient that a higher score on the scale means better functioning) [35] |
PROMs: patient reported outcome measures
*’Better’ is defined as higher scores indicating “better” outcome
**’Normed’ is defined as normed to the general U.S. population
***’More’ is defined as higher scores indicating “more” of what was being measured
How to distinguish clinically relevant PROMs scores
Clinically alarming scores were put in perspective by comparing current scores to the patients’ previous scores or to norm population scores. Ten studies described strategies to distinguish clinically relevant scores in practice [3–5, 16–18, 35]. These studies were conducted internationally among patients with different cancer diagnoses. Most frequently used comparator scores were the patients’ own previous PROMs data [3–5, 16–18, 35]. Additionally, four studies used a norm population (i.e.; patients with the same disease) to determine the clinical relevance of scores [1, 5, 19, 37]. Patients pointed out that the comparison of their own data with the scores from ‘patients like them’ (i.e. same sex, and age) was most valuable: it puts the patients functioning in perspective of what is regarded ‘normal’ [5].
Studies showed different methods of applying color to highlight scores, including: visualizing clinically non-alarming scores in green and clinically alarming scores in red [17]; red and green shading to show undesirable and desirable score ranges, respectively [4, 5, 18, 25]; a background fading from red (bottom) to green (top) [1]; traffic light colors (i.e. green, orange, red) [2].
Smith et al. [18] reported that 74% of patients and 80% of clinicians preferred green shading of non-alarming scores or red shading of alarming scores. Hildon et al. [2] reported that traffic light colors could improve interpretation accuracy across different graphic visualization formats, since these are universally recognized. Loth et al. [17] reported that 93% of the studied patients correctly interpreted the meaning of traffic light colors. Nonetheless, PROMs visualization strategies must include a detailed explanatory legend of the meaning and interpretation of colors and scores [4].
Other visualization strategies included the use of red circles around important scores—these were interpreted more accurately than green shading [34]. Furthermore, threshold lines across score bars were used to indicate whether scores are better or worse than threshold scores. These visualization techniques pleased 69% of patients and 70% of clinicians [17, 18]. Contrastingly, Snyder et al. [34] reported that a threshold-line was significantly more often rated as ‘very clear’, and most useful compared to green shading, and red circles. In more detail, another study found that a dotted threshold line was preferred over a solid threshold line to indicate alarming scores in bar charts [5]. Lastly, exclamation points can be used to indicate possibly concerning score changes, which was the preferred method for 79% of 39 patients and 40% of 40 clinicians [18].
Discussion
This systematic review included 25 studies in different healthcare settings throughout Europe, the US, and Canada, that reported about preferences and interpretation accuracy of patients and clinicians for the visualization of PROMs scores. Very few graphical visualization formats for presenting PROMs data could be identified. Overall, a limited amount of literature was found on this matter, which was summarized as follows. For individual level PROMs data, patients and clinicians preferred line graphs and bar charts, since they were considered visually clear. Bar charts were most often interpreted correctly by patients, while clinicians had high interpretation accuracy across all graph formats. For presenting group level PROMs data, pie charts and line graphs were preferred: patients most often interpreted line graphs correctly; for proportional data, clinicians most often accurately interpreted pie charts. To guide clinical interpretation by distinguishing clinically relevant scores, PROMs scores were most often compared to patients’ previous scores, followed by comparison to mean norm population scores. Here, correct interpretation can be supported by highlighting patients’ clinically alarming scores with colors, and by using threshold lines across score bars or lines, or circles around alarming scores.
Furthermore, we looked into the challenges that may hinder graphic visualization format interpretation. An underlying cause of incorrect interpretation of graphs may be the lack of standardization in rules for interpretation; variability exists in score directionality (e.g. higher scores can either indicate better or worse outcomes), and scaling (e.g. scores ranging from 0 to 100 indicating the worst-to-best possible scores, or scores ‘normed’ to a defined population). Furthermore, meaningful interpretation of PROMs scores is complicated by the way the statistical and clinical significance of the findings (i.e. thresholds to distinguish clinical importance) are addressed [16, 18, 38, 39]. Therefore, effort must be made to present PROMs scores to patients and clinicians more accurately by: improving directional consistency by making use of standard descriptive labels [34, 37], clear label ratings (i.e. consistent scales ranging from 0–100 [37]), detailed information on the meaning of high and low scores [17], simple iconography [25], and brief definitions to understand what the PROMs scores represent [25]. Furthermore, it was suggested is to visualize only a limited number of symptoms, and to ask patients to prioritize the symptoms they want feedback on [35]. Afore mentioned factors to guide clinical interpretation of graphic visualization formats of PROMs data, were identified as well in a Delphi-consensus study by Snyder et al. [15]. In this study, a panel including 15 doctor or nurse clinicians, 10 participants who identified as patient or caregiver advocates, 12 researchers, and 6 members of journal editorial boards, were asked to review data display issues, and give their perspectives on these issues to develop consensus statements. The authors conclude that implementation of graphic visualization formats of PROMs data have enormous potential to promote patient-centred care, however, it is critical that patients and clinicians understand what PROMs scores mean. More specifically, they recommended to use exceptionally clear labelling, titling, and other annotations to address potential confusion in direction of scores, and warn for mixing score direction in a single display. Furthermore, for conveying score meaning, descriptive labels along the y-axis are expected to be helpful [15]. The Setting International Standards of Patient-Reported Outcomes and Quality of Life Endpoints in Cancer Clinical Trials (SISAQoL) guideline [40] reported recommendations on directionality of scale scores similar to Snyder et al. [15]. In order to enhance clinicians’ interpretation of PROMs scores they recommended to reduce the number of metrics presented (e.g. a maximum of six bars in bar graphs and 4 lines in line graphs), use coloured arrows (e.g. green for better and red for worse scores) and to accompany more complex displays like funnel plots with a detailed interpretation [40]. This guideline will contribute to standardize rules for interpretation and visualization.
Another suggestion for clarifying PROMs visualisation formats is to develop a dynamic dashboard for PROMs feedback [3, 6, 28]. This way, patients and clinicians are able to change between different graphic visualization formats. We imagine options such as comparing scores to norms or threshold scores, as not all patients may want to do so. Based on the hypothesis that serving individual preferences may facilitate interpretation accuracy, this could improve interpretation of PROMs scores as well. However, it should be taken into account that implementing a dynamic dashboard comes with challenges, like access and availability of suitable software and sufficient IT staff to support such a dashboard.
Last, the implementation of PROMs hinges on more factors than visualization of data, starting with the perceived value by patients and clinicians of discussing PROMs during clinical encounters. Nine studies included in our review noted that PROMs were perceived as valuable [4, 5, 12, 16, 25–27, 35, 36], where some studies (n = 5) showed mixed results regarding the usefulness of PROMs [1, 3, 14, 17, 19]. How patients value PROMs may determine the interest in graphic visualization of PROMs. This could have affected the presented results in this review.
This systematic review is limited by the fact that only papers published in English were included. Studies in other languages regarding locally successful implemented feedback of PROMs data might have been missed. Furthermore, in this review, no distinction is made between objective (i.e. does a person actually interpret scores correctly), and subjective (i.e. does the person says he or she interprets scores correctly) interpretation accuracy. Only two included studies made a distinction between the different types of interpretation accuracy [4, 17]; for the other studies, based on the description of the study methods, we believe objective interpretation accuracy was assessed. Therefore, future research may investigate potential differences between actual and perceived interpretation accuracy.
Another potential bias is that for the majority of included studies it remains unclear how questions about preferences and interpretation of the presented visualization format(s) were framed to the study population. Besides, study participants may have had different background knowledge about graphic visualization formats, presentation of formats, and the content of PROMs data. This could confound our findings about how data and graphs were interpret, understood, and valued.
Furthermore, this review did not consider different levels of patients’ health literacy, since this information was not presented in the majority of articles reviewed. Patients with a higher level of health literacy can have different preferences compared to patients with lower levels of health literacy [3]. Some patients with limited health literacy may not understand the longitudinal nature of data from left to right nor the temporal connection between different graphical elements [28]. Therefore, healthy literacy should be included as factor in studies dealing with data interpretation, for example measured through the use of the validated Health Literacy Questionnaire (HLQ) [41]. Furthermore, health literacy can differ among age, gender, and/or education level and therefore should be studied among these separate groups of patients in relation to PROMs visualization [42, 43].
In conclusion, there was no predominant graphical visualization format approach in terms of preferences or interpretation accuracy for both patients and clinicians. To guide the clinical interpretation of scores during clinical encounters, PROMs scores can be compared to patients’ previous scores or mean scores from a norm population, or compared to comparator scores or score thresholds. Furthermore, the use of colors, threshold lines, or circles around alarming scores can visualize the clinical meaning of PROMs scores that have been compared to previous scores, norms, thresholds, or reference populations. Finally, detailed clarification of each graph may be essential for accurate interpretation. Visualization strategies should therefore include descriptions of PROMs directionality as well as standard descriptive labels, brief definitions, and presentation of a limited number of symptoms in a graph.
Supplementary Information
Additional file 1: Table S1. Search strategy for MEDLINE (accessed through PubMed), Embase (accessed through Ovid Platform), PsycINFO (accessed through Ovid Platform) and CINAHL.
Acknowledgements
Not applicable.
Abbreviations
- CASP
Critical appraisal skills programme
- CDAI
Clinical disease activity index
- HER
Electronic Health Record
- EORTC-QLQ-C30
European organization for research and treatment of cancer quality of life questionnaire Core 30
- HADS
Hospital anxiety and depression scale
- HCD
Human centered design
- HRQoL
Health-related quality of life
- HUI
Health utility index
- Health-ITUES
Health IT usability evaluation scale
- HLQ
Health literacy questionnaire
- JHCRN
Johns Hopkins clinical research network
- N/A
Not applicable
- PRISMA
Preferred reporting items for systematic reviews and meta-analysis
- PROMs
Patient reported outcome measurements
- PROMIS
Patient-reported outcomes measurement information system
- QoL
Quality of life
- REALM-SF
Rapid estimate of adult literacy in medicine short form
- SISAQoL
Setting international standards of patient-reported outcomes and quality of life endpoints in cancer clinical trials
- SD
Standard deviation
- USA
United States of America
- WHOQOL
World health organization quality of life
Authors' contributions
EA, IF: Conceptualization, article screening, data analysis, editing, writing. KdL: conceptualization, literature search, data analysis, editing, reviewing, writing. IW, LvdP: conceptualization, editing, reviewing, writing. EW: literature search, reviewing method section, PRISMA. SS, MW, IvdP: conceptualization, reviewing, writing. All authors read and approved the final manuscript.
Funding
Not applicable.
Availability of data and materials
Data sharing is not applicable to this article as no datasets were generated or analysed during the current study.
Declarations
Ethics approval and consent to participate
Not applicable.
Consent for publication
Not applicable.
Competing interests
The authors declare that they have no competing interests.
Footnotes
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Elaine A. C. Albers and Itske Fraterman have contributed equally to this work
References
- 1.Fischer KI, De Faoite D, Rose M. Patient-reported outcomes feedback report for knee arthroplasty patients should present selective information in a simple design - findings of a qualitative study. J Patient Rep Outcomes. 2020;4(1):6. doi: 10.1186/s41687-020-0173-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Hildon Z, Allwood D, Black N. Making data more meaningful: patients' views of the format and content of quality indicators comparing health care providers. Patient Educ Couns. 2012;88(2):298–304. doi: 10.1016/j.pec.2012.02.006. [DOI] [PubMed] [Google Scholar]
- 3.Izard J, Hartzler A, Avery DI, Shih C, Dalkin BL, Gore JL. User-centered design of quality of life reports for clinical care of patients with prostate cancer. Surgery. 2014;155(5):789–796. doi: 10.1016/j.surg.2013.12.007. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Kuijpers W, Giesinger JM, Zabernigg A, Young T, Friend E, Tomaszewska IM, et al. Patients' and health professionals' understanding of and preferences for graphical presentation styles for individual-level EORTC QLQ-C30 scores. Qual Life Res. 2016;25(3):595–604. doi: 10.1007/s11136-015-1107-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Oerlemans S, Arts LP, Horevoorts NJ, van de Poll-Franse LV (2017) "Am I normal?" The Wishes of patients with lymphoma to compare their patient-reported outcomes with those of their peers. J Med Internet Res. 19(8):e288. [DOI] [PMC free article] [PubMed]
- 6.Hartzler AL, Izard JP, Dalkin BL, Mikles SP, Gore JL. Design and feasibility of integrating personalized PRO dashboards into prostate cancer care. J Am Med Inform Assoc. 2016;23(1):38–47. doi: 10.1093/jamia/ocv101. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Snyder C, Brundage M, Rivera YM, Wu AW (2019) A PRO-cision medicine methods toolkit to address the challenges of personalizing cancer care using patient-reported outcomes: introduction to the supplement. Med Care. 57:S1-s7 [DOI] [PMC free article] [PubMed]
- 8.Snyder CF, Blackford AL, Aaronson NK, Detmar SB, Carducci MA, Brundage MD, et al. Can patient-reported outcome measures identify cancer patients' most bothersome issues? J Clin Oncol. 2011;29(9):1216–1220. doi: 10.1200/JCO.2010.33.2080. [DOI] [PubMed] [Google Scholar]
- 9.Berry DL, Blumenstein BA, Halpenny B, Wolpin S, Fann JR, Austin-Seymour M, et al. Enhancing patient-provider communication with the electronic self-report assessment for cancer: a randomized trial. J Clin Oncol. 2011;29(8):1029–1035. doi: 10.1200/JCO.2010.30.3909. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Yang LY, Manhas DS, Howard AF, Olson RA. Patient-reported outcome use in oncology: a systematic review of the impact on patient-clinician communication. Supp Care Cancer. 2018;26(1):41–60. doi: 10.1007/s00520-017-3865-7. [DOI] [PubMed] [Google Scholar]
- 11.Chen J, Ou L, Hollis SJ. A systematic review of the impact of routine collection of patient reported outcome measures on patients, providers and health organisations in an oncologic setting. BMC Health Serv Res. 2013;13:211. doi: 10.1186/1472-6963-13-211. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Brundage M, Leis A, Bezjak A, Feldman-Stewart D, Degner L, Velji K, et al. Cancer patients' preferences for communicating clinical trial quality of life information: a qualitative study. Qual Life Res. 2003;12(4):395–404. doi: 10.1023/A:1023404731041. [DOI] [PubMed] [Google Scholar]
- 13.Bantug ET, Coles T, Smith KC, Snyder CF, Rouette J, Brundage MD, et al. Graphical displays of patient-reported outcomes (PRO) for use in clinical practice: What makes a pro picture worth a thousand words? Patient Educ Couns. 2016;99(4):483–490. doi: 10.1016/j.pec.2015.10.027. [DOI] [PubMed] [Google Scholar]
- 14.Damman OC, Verbiest MEA, Vonk SI, Berendse HW, Bloem BR, de Bruijne MC, et al. Using PROMs during routine medical consultations: The perspectives of people with Parkinson's disease and their health professionals. Health Expect. 2019;22(5):939–951. doi: 10.1111/hex.12899. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Snyder C, Smith K, Holzner B, Rivera YM, Bantug E, Brundage M, et al. Making a picture worth a thousand numbers: recommendations for graphically displaying patient-reported outcomes data. Qual Life Res. 2019;28(2):345–356. doi: 10.1007/s11136-018-2020-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Brundage MD, Smith KC, Little EA, Bantug ET, Snyder CF. Communicating patient-reported outcome scores using graphic formats: results from a mixed-methods evaluation. Qual Life Res. 2015;24(10):2457–2472. doi: 10.1007/s11136-015-0974-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Loth FL, Holzner B, Sztankay M, Bliem HR, Raoufi S, Rumpold G, et al. Cancer patients' understanding of longitudinal EORTC QLQ-C30 scores presented as bar charts. Patient Educ Couns. 2016;99(12):2012–2017. doi: 10.1016/j.pec.2016.08.004. [DOI] [PubMed] [Google Scholar]
- 18.Smith KC, Brundage MD, Tolbert E, Little EA, Bantug ET, Snyder CF, et al. Engaging stakeholders to improve presentation of patient-reported outcomes data in clinical practice. Supp Care Cancer. 2016;24(10):4149–4157. doi: 10.1007/s00520-016-3240-0. [DOI] [PubMed] [Google Scholar]
- 19.van Overveld LFJ, Takes RP, Vijn TW, Braspenning JCC, de Boer JP, Brouns JJA, et al. Feedback preferences of patients, professionals and health insurers in integrated head and neck cancer care. Health Expect. 2017;20(6):1275–1288. doi: 10.1111/hex.12567. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Tolbert E, Brundage M, Bantug E, Blackford AL, Smith K, Snyder C, et al. In proportion: approaches for displaying patient-reported outcome research study results as percentages responding to treatment. Qual Life Res. 2019;28(3):609–620. doi: 10.1007/s11136-018-2065-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Basch E, Barbera L, Kerrigan CL, Velikova G. Implementation of patient-reported outcomes in routine medical care. Am Soc Clin Oncol Educ Book. 2018;38:122–134. doi: 10.1200/EDBK_200383. [DOI] [PubMed] [Google Scholar]
- 22.Shamseer L, Moher D, Clarke M, Ghersi D, Liberati A, Petticrew M, et al (2015) Preferred reporting items for systematic review and meta-analysis protocols (PRISMA-P) 2015: elaboration and explanation. Bmj.350:g7647 [DOI] [PubMed]
- 23.Bramer WM, Giustini D, de Jonge GB, Holland L, Bekhuis T. De-duplication of database search results for systematic reviews in EndNote. J Med Libr Assoc. 2016;104(3):240–243. doi: 10.3163/1536-5050.104.3.014. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Nadelson S, Nadelson LS. Evidence-based practice article reviews using CASP tools: a method for teaching EBP. Worldviews Evid Based Nurs. 2014;11(5):344–346. doi: 10.1111/wvn.12059. [DOI] [PubMed] [Google Scholar]
- 25.Ragouzeos D, Gandrup J, Berrean B, Li J, Murphy M, Trupin L, et al. "Am I OK?" using human centered design to empower rheumatoid arthritis patients through patient reported outcomes. Patient Educ Couns. 2019;102(3):503–510. doi: 10.1016/j.pec.2018.10.016. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Geerards D, Pusic A, Hoogbergen M, van der Hulst R, Sidey-Gibbons C (2019) Computerized Quality of Life Assessment: A Randomized Experiment to Determine the Impact of Individualized Feedback on Assessment Experience. J Med Internet Res.21(7):e12212. [DOI] [PMC free article] [PubMed]
- 27.Grossman LV, Feiner SK, Mitchell EG, Masterson Creber RM. Leveraging patient-reported outcomes using data visualization. Appl Clin Inform. 2018;9(3):565–575. doi: 10.1055/s-0038-1667041. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Liu LH, Garrett SB, Li J, Ragouzeos D, Berrean B, Dohan D, et al. Patient and clinician perspectives on a patient-facing dashboard that visualizes patient reported outcomes in rheumatoid arthritis. Health Expect. 2020;23(4):846–859. doi: 10.1111/hex.13057. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Tolbert E, Brundage M, Bantug E, Blackford AL, Smith K, Snyder C, et al. Picture this: presenting longitudinal patient-reported outcome research study results to patients. Med Decis Making. 2018;38(8):994–1005. doi: 10.1177/0272989X18791177. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Brundage M, Feldman-Stewart D, Leis A, Bezjak A, Degner L, Velji K, et al. Communicating quality of life information to cancer patients: a study of six presentation formats. J Clin Oncol. 2005;23(28):6949–6956. doi: 10.1200/JCO.2005.12.514. [DOI] [PubMed] [Google Scholar]
- 31.Brundage MD, Smith KC, Little EA, Bantug ET, Snyder CF (2015) Board PRODPSA. Communicating patient-reported outcome scores using graphic formats: results from a mixed-methods evaluation. Qual Life Res. 24(10):2457–72 [DOI] [PMC free article] [PubMed]
- 32.McNair AG, Brookes ST, Davis CR, Argyropoulos M, Blazeby JM. Communicating the results of randomized clinical trials: do patients understand multidimensional patient-reported outcomes? J Clin Oncol. 2010;28(5):738–743. doi: 10.1200/JCO.2009.23.9111. [DOI] [PubMed] [Google Scholar]
- 33.Santana MJ, Feeny DH. Using the health utilities index in routine clinical care: process, feasibility, and acceptability: a randomized controlled trial. Patient. 2009;2(3):159–167. doi: 10.2165/11313620-000000000-00000. [DOI] [PubMed] [Google Scholar]
- 34.Snyder CF, Smith KC, Bantug ET, Tolbert EE, Blackford AL, Brundage MD, et al. What do these scores mean? Presenting patient-reported outcomes data to patients and clinicians to improve interpretability. Cancer. 2017;123(10):1848–1859. doi: 10.1002/cncr.30530. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Jagsi R, Chiang A, Polite BN, Medeiros BC, McNiff K, Abernethy AP, et al. Qualitative analysis of practicing oncologists' attitudes and experiences regarding collection of patient-reported outcomes. J Oncol Pract. 2013;9(6):e290–e297. doi: 10.1200/JOP.2012.000823. [DOI] [PubMed] [Google Scholar]
- 36.Hartzler AL, Chaudhuri S, Fey BC, Flum DR, Lavallee D. Integrating Patient-Reported Outcomes into Spine Surgical Care through Visual Dashboards: Lessons Learned from Human-Centered Design. EGEMS (Wash DC) 2015;3(2):1133. doi: 10.13063/2327-9214.1133. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Brundage M, Blackford A, Tolbert E, Smith K, Bantug E, Snyder C, et al. Presenting comparative study PRO results to clinicians and researchers: beyond the eye of the beholder. Qual Life Res. 2018;27(1):75–90. doi: 10.1007/s11136-017-1710-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Brundage M, Bass B, Jolie R, Foley K. A knowledge translation challenge: clinical use of quality of life data from cancer clinical trials. Qual Life Res. 2011;20(7):979–985. doi: 10.1007/s11136-011-9848-0. [DOI] [PubMed] [Google Scholar]
- 39.Snyder CF, Aaronson NK. Use of patient-reported outcomes in clinical practice. Lancet. 2009;374(9687):369–370. doi: 10.1016/S0140-6736(09)61400-8. [DOI] [PubMed] [Google Scholar]
- 40.Hancock SL, Ryan OF, Marion V, Kramer S, Kelly P, Breen S, et al (2020) Feedback of patient-reported outcomes to healthcare professionals for comparing health service performance: a scoping review. BMJ Open. 10(11):e038190 [DOI] [PMC free article] [PubMed]
- 41.Osborne RH, Batterham RW, Elsworth GR, Hawkins M, Buchbinder R. The grounded psychometric development and initial validation of the Health Literacy Questionnaire (HLQ) BMC Public Health. 2013;13:658. doi: 10.1186/1471-2458-13-658. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Williams MV, Davis T, Parker RM, Weiss BD. The role of health literacy in patient-physician communication. Fam Med. 2002;34(5):383–389. [PubMed] [Google Scholar]
- 43.Clouston SAP, Manganello JA, Richards M. A life course approach to health literacy: the role of gender, educational attainment and lifetime cognitive capability. Age Ageing. 2017;46(3):493–499. doi: 10.1093/ageing/afw229. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Additional file 1: Table S1. Search strategy for MEDLINE (accessed through PubMed), Embase (accessed through Ovid Platform), PsycINFO (accessed through Ovid Platform) and CINAHL.
Data Availability Statement
Data sharing is not applicable to this article as no datasets were generated or analysed during the current study.